atomic_read 654 arch/alpha/kernel/smp.c if (atomic_read(&mm->mm_users) <= 1) { atomic_read 701 arch/alpha/kernel/smp.c if (atomic_read(&mm->mm_users) <= 1) { atomic_read 755 arch/alpha/kernel/smp.c if (atomic_read(&mm->mm_users) <= 1) { atomic_read 301 arch/arc/mm/tlb.c if (atomic_read(&mm->mm_users) == 0) atomic_read 188 arch/arm/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 112 arch/arm/kernel/machine_kexec.c while ((atomic_read(&waiting_for_crash_ipi) > 0) && msecs) { atomic_read 116 arch/arm/kernel/machine_kexec.c if (atomic_read(&waiting_for_crash_ipi) > 0) atomic_read 155 arch/arm/kvm/coproc.c ncores = atomic_read(&vcpu->kvm->online_vcpus) - 1; atomic_read 172 arch/arm/kvm/coproc.c if (atomic_read(&vcpu->kvm->online_vcpus) > 1) atomic_read 210 arch/arm/mach-exynos/pm.c if (atomic_read(&cpu1_wakeup)) atomic_read 253 arch/arm/mach-exynos/pm.c !atomic_read(&cpu1_wakeup)) atomic_read 256 arch/arm/mach-exynos/pm.c if (!atomic_read(&cpu1_wakeup)) atomic_read 260 arch/arm/mach-exynos/pm.c while (!atomic_read(&cpu1_wakeup)) { atomic_read 255 arch/arm64/include/asm/pgtable.h if (mm != current->active_mm && atomic_read(&mm->mm_users) <= 1) atomic_read 232 arch/arm64/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 204 arch/arm64/kernel/insn.c while (atomic_read(&pp->cpu_count) <= num_online_cpus()) atomic_read 1031 arch/arm64/kernel/smp.c while ((atomic_read(&waiting_for_crash_ipi) > 0) && timeout--) atomic_read 1034 arch/arm64/kernel/smp.c if (atomic_read(&waiting_for_crash_ipi) > 0) atomic_read 1043 arch/arm64/kernel/smp.c return (atomic_read(&waiting_for_crash_ipi) > 0); atomic_read 145 arch/csky/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 157 arch/hexagon/kernel/kgdb.c if (atomic_read(&kgdb_active) != -1) { atomic_read 40 arch/ia64/include/asm/atomic.h old = atomic_read(v); \ atomic_read 55 arch/ia64/include/asm/atomic.h old = atomic_read(v); \ atomic_read 85 arch/ia64/include/asm/tlbflush.h if (atomic_read(&mm->mm_users) == 0) atomic_read 70 arch/ia64/kernel/crash.c if (atomic_read(&kdump_cpu_frozen) == cpu_num) atomic_read 161 arch/ia64/kernel/crash.c if (atomic_read(&kdump_in_progress)) { atomic_read 49 arch/ia64/kernel/irq.c seq_printf(p, "ERR: %10u\n", atomic_read(&irq_err_count)); atomic_read 383 arch/ia64/kernel/kprobes.c i = atomic_read(&kcb->prev_kprobe_index); atomic_read 1604 arch/ia64/kernel/mca.c if (atomic_read(&kdump_in_progress)) atomic_read 1707 arch/ia64/kernel/mca.c while (monarch_cpu == -1 && !atomic_read(&kdump_in_progress)) atomic_read 1718 arch/ia64/kernel/mca.c while (monarch_cpu != -1 && !atomic_read(&kdump_in_progress)) atomic_read 298 arch/ia64/kernel/smp.c if (likely(mm == current->active_mm && atomic_read(&mm->mm_users) == 1)) atomic_read 124 arch/ia64/kernel/uncached.c if (atomic_read(&uc_pool->status)) atomic_read 143 arch/ia64/kernel/uncached.c if (atomic_read(&uc_pool->status)) atomic_read 408 arch/ia64/pci/pci.c BUG_ON(atomic_read(&dev->enable_cnt)); atomic_read 118 arch/m68k/amiga/chipram.c unsigned long n = atomic_read(&chipavail); atomic_read 53 arch/m68k/include/asm/atomic.h : "g" (i), "2" (atomic_read(v))); \ atomic_read 68 arch/m68k/include/asm/atomic.h : "g" (i), "2" (atomic_read(v))); \ atomic_read 176 arch/m68k/include/asm/atomic.h prev = atomic_read(v); atomic_read 189 arch/m68k/include/asm/atomic.h prev = atomic_read(v); atomic_read 37 arch/m68k/kernel/irq.c seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); atomic_read 32 arch/microblaze/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 94 arch/mips/cavium-octeon/setup.c while (!atomic_read(&kexec_ready_to_reboot)) atomic_read 47 arch/mips/kernel/crash.c while (!atomic_read(&kexec_ready_to_reboot)) atomic_read 332 arch/mips/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 43 arch/mips/kernel/irq.c seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); atomic_read 324 arch/mips/kernel/kgdb.c if (atomic_read(&kgdb_active) != -1) atomic_read 332 arch/mips/kernel/kgdb.c if (atomic_read(&kgdb_setting_breakpoint)) atomic_read 126 arch/mips/kernel/machine_kexec.c while (!atomic_read(&kexec_ready_to_reboot)) atomic_read 632 arch/mips/kernel/perf_event_mipsxx.c if (atomic_read(&active_events) == 0) atomic_read 98 arch/mips/kernel/pm-cps.c while (atomic_read(a) < online) atomic_read 106 arch/mips/kernel/pm-cps.c while (atomic_read(a) > online) atomic_read 538 arch/mips/kernel/smp.c } else if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) { atomic_read 588 arch/mips/kernel/smp.c } else if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) { atomic_read 655 arch/mips/kernel/smp.c } else if ((atomic_read(&vma->vm_mm->mm_users) != 1) || atomic_read 67 arch/mips/kernel/spinlock_test.c while (atomic_read(&s->enter_wait)) atomic_read 74 arch/mips/kernel/spinlock_test.c while (atomic_read(&s->start_wait)) atomic_read 86 arch/mips/kernel/spinlock_test.c while (atomic_read(&s->exit_wait)) atomic_read 50 arch/mips/kernel/sync-r4k.c while (atomic_read(&count_count_start) != 1) atomic_read 71 arch/mips/kernel/sync-r4k.c while (atomic_read(&count_count_stop) != 1) atomic_read 101 arch/mips/kernel/sync-r4k.c while (atomic_read(&count_count_start) != 2) atomic_read 111 arch/mips/kernel/sync-r4k.c while (atomic_read(&count_count_stop) != 2) atomic_read 214 arch/mips/kernel/traps.c } else if (atomic_read(&kgdb_active) != -1 && atomic_read 164 arch/mips/kvm/mips.c for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) atomic_read 183 arch/mips/math-emu/dsemul.c fr_idx = atomic_read(¤t->thread.bd_emu_frame); atomic_read 249 arch/mips/math-emu/dsemul.c fr_idx = atomic_read(¤t->thread.bd_emu_frame); atomic_read 237 arch/mips/sgi-ip27/ip27-nmi.c while (atomic_read(&nmied_cpus) != num_online_cpus()); atomic_read 215 arch/nds32/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 871 arch/nds32/kernel/perf_event_cpu.c if (atomic_read(active_events) == 0) { atomic_read 53 arch/openrisc/kernel/sync-timer.c while (atomic_read(&count_count_start) != 1) atomic_read 74 arch/openrisc/kernel/sync-timer.c while (atomic_read(&count_count_stop) != 1) atomic_read 104 arch/openrisc/kernel/sync-timer.c while (atomic_read(&count_count_start) != 2) atomic_read 114 arch/openrisc/kernel/sync-timer.c while (atomic_read(&count_count_stop) != 2) atomic_read 25 arch/parisc/include/asm/mmu_context.h BUG_ON(atomic_read(&mm->mm_users) != 1); atomic_read 40 arch/parisc/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 146 arch/powerpc/include/asm/book3s/64/tlbflush.h if (atomic_read(&vma->vm_mm->context.copros) > 0) atomic_read 114 arch/powerpc/include/asm/eeh.h return pe ? !!atomic_read(&pe->pass_dev_cnt) : false; atomic_read 526 arch/powerpc/include/asm/kvm_book3s_64.h if (atomic_read(&kvm->arch.hpte_mod_interest)) atomic_read 15 arch/powerpc/include/asm/membarrier.h if (likely(!(atomic_read(&next->membarrier_state) & atomic_read 66 arch/powerpc/include/asm/tlb.h if (atomic_read(&mm->context.active_cpus) > 1) atomic_read 72 arch/powerpc/include/asm/tlb.h WARN_ON(atomic_read(&mm->context.copros) > 0); atomic_read 126 arch/powerpc/kernel/crash.c while ((atomic_read(&cpus_in_crash) < ncpus) && (--msecs > 0)) atomic_read 131 arch/powerpc/kernel/crash.c if (atomic_read(&cpus_in_crash) >= ncpus) { atomic_read 137 arch/powerpc/kernel/crash.c ncpus - atomic_read(&cpus_in_crash)); atomic_read 168 arch/powerpc/kernel/crash.c while (atomic_read(&cpus_in_crash) < ncpus) atomic_read 754 arch/powerpc/kernel/rtas.c while (rc == H_MULTI_THREADS_ACTIVE && !atomic_read(&data->done) && atomic_read 755 arch/powerpc/kernel/rtas.c !atomic_read(&data->error)) atomic_read 758 arch/powerpc/kernel/rtas.c if (rc || atomic_read(&data->error)) { atomic_read 763 arch/powerpc/kernel/rtas.c if (atomic_read(&data->error)) atomic_read 764 arch/powerpc/kernel/rtas.c rc = atomic_read(&data->error); atomic_read 800 arch/powerpc/kernel/rtas.c while (rc == H_SUCCESS && !atomic_read(&data->done) && !atomic_read(&data->error)) atomic_read 1000 arch/powerpc/kernel/rtas.c if (atomic_read(&data.error) != 0) atomic_read 1015 arch/powerpc/kernel/rtas.c return atomic_read(&data.error); atomic_read 383 arch/powerpc/kernel/smp.c spin_until_cond(atomic_read(&__nmi_ipi_lock) == 0); atomic_read 392 arch/powerpc/kernel/smp.c spin_until_cond(atomic_read(&__nmi_ipi_lock) == 0); atomic_read 398 arch/powerpc/kernel/smp.c WARN_ON(atomic_read(&__nmi_ipi_lock) != 1); atomic_read 602 arch/powerpc/kernel/time.c if (atomic_read(&ppc_n_lost_interrupts) != 0) atomic_read 955 arch/powerpc/kernel/trace/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 136 arch/powerpc/kvm/book3s_64_mmu_hv.c if (atomic_read(&kvm->arch.vcpus_running)) { atomic_read 1041 arch/powerpc/kvm/book3s_64_mmu_hv.c return atomic_read(&kvm->arch.vcpus_running) != 0; atomic_read 1866 arch/powerpc/kvm/book3s_64_mmu_hv.c if (atomic_read(&kvm->arch.vcpus_running)) { atomic_read 3230 arch/powerpc/kvm/book3s_hv.c int n_online = atomic_read(&vc->online_count); atomic_read 5323 arch/powerpc/kvm/book3s_hv.c if (atomic_read(&kvm->arch.vcpus_running)) { atomic_read 184 arch/powerpc/kvm/book3s_hv_builtin.c return atomic_read(&hv_vm_count) != 0; atomic_read 200 arch/powerpc/kvm/book3s_xive_template.c if (atomic_read(&q->pending_count)) { atomic_read 204 arch/powerpc/kvm/book3s_xive_template.c WARN_ON(p > atomic_read(&q->count)); atomic_read 479 arch/powerpc/kvm/powerpc.c for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) atomic_read 344 arch/powerpc/mm/book3s64/pgtable.c BUG_ON(atomic_read(&page->pt_frag_refcount) <= 0); atomic_read 106 arch/powerpc/mm/book3s64/radix_hugetlbpage.c (atomic_read(&mm->context.copros) > 0)) atomic_read 1044 arch/powerpc/mm/book3s64/radix_pgtable.c if ((change & _PAGE_RW) && atomic_read(&mm->context.copros) > 0) { atomic_read 1078 arch/powerpc/mm/book3s64/radix_pgtable.c (atomic_read(&mm->context.copros) > 0)) atomic_read 364 arch/powerpc/mm/book3s64/radix_tlb.c if (atomic_read(&mm->context.copros) > 0) atomic_read 497 arch/powerpc/mm/book3s64/radix_tlb.c if (atomic_read(&mm->context.copros) > 0) atomic_read 551 arch/powerpc/mm/book3s64/radix_tlb.c if (atomic_read(&mm->context.copros) > 0) atomic_read 617 arch/powerpc/mm/book3s64/radix_tlb.c if (atomic_read(&mm->context.copros) > 0) atomic_read 619 arch/powerpc/mm/book3s64/radix_tlb.c if (atomic_read(&mm->mm_users) <= 1 && current->mm == mm) atomic_read 631 arch/powerpc/mm/book3s64/radix_tlb.c if (atomic_read(&mm->context.copros) > 0) atomic_read 279 arch/powerpc/mm/hugetlbpage.c if (atomic_read(&tlb->mm->mm_users) < 2 || atomic_read 113 arch/powerpc/mm/pgtable-frag.c BUG_ON(atomic_read(&page->pt_frag_refcount) <= 0); atomic_read 41 arch/powerpc/perf/8xx-pmu.c ctr = atomic_read(&instruction_counter); atomic_read 43 arch/powerpc/perf/8xx-pmu.c } while (ctr != atomic_read(&instruction_counter)); atomic_read 1990 arch/powerpc/perf/core-book3s.c if (atomic_read(&num_events) == 0 && atomic_read 222 arch/powerpc/perf/core-fsl-emb.c if (atomic_read(&num_events)) { atomic_read 571 arch/powerpc/perf/core-fsl-emb.c if (atomic_read(&num_events) == 0 && atomic_read 441 arch/powerpc/platforms/85xx/smp.c while ( (atomic_read(&kexec_down_cpus) != (num_cpus - 1)) && atomic_read 37 arch/powerpc/platforms/cell/cpufreq_spudemand.c busy_spus = atomic_read(&cbe_spu_info[cpu_to_node(cpu)].busy_spus); atomic_read 1459 arch/powerpc/platforms/cell/spufs/file.c if (atomic_read(&inode->i_count) != 1) atomic_read 354 arch/powerpc/platforms/cell/spufs/inode.c if ((cbe_spu_info[node].n_spus - atomic_read( atomic_read 405 arch/powerpc/platforms/cell/spufs/sched.c if (atomic_read(&ctx->gang->aff_sched_count) == 0) atomic_read 1079 arch/powerpc/platforms/cell/spufs/sched.c atomic_read(&nr_spu_contexts), atomic_read 103 arch/powerpc/platforms/powermac/backlight.c if (atomic_read(&kernel_backlight_disabled)) atomic_read 131 arch/powerpc/platforms/powermac/backlight.c if (atomic_read(&kernel_backlight_disabled)) atomic_read 170 arch/powerpc/platforms/powermac/backlight.c if (atomic_read(&kernel_backlight_disabled)) atomic_read 178 arch/powerpc/platforms/powermac/backlight.c if (atomic_read(&kernel_backlight_disabled)) atomic_read 645 arch/powerpc/platforms/powernv/idle.c if (atomic_read(&local_paca->dont_stop)) { atomic_read 1230 arch/powerpc/platforms/powernv/vas-window.c if (!window->tx_win && atomic_read(&window->num_txwins) != 0) { atomic_read 64 arch/powerpc/platforms/pseries/suspend.c if (atomic_read(&suspending)) atomic_read 944 arch/powerpc/platforms/pseries/vio.c return sprintf(buf, "%d\n", atomic_read(&viodev->cmo.allocs_failed)); atomic_read 168 arch/powerpc/sysdev/xive/common.c if (atomic_read(&q->pending_count)) { atomic_read 171 arch/powerpc/sysdev/xive/common.c WARN_ON(p > atomic_read(&q->count)); atomic_read 882 arch/powerpc/xmon/xmon.c if (!bp->enabled && atomic_read(&bp->ref_count) == 0) { atomic_read 136 arch/riscv/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 116 arch/s390/include/asm/mmu_context.h while (atomic_read(&mm->context.flush_count)) atomic_read 209 arch/s390/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 277 arch/s390/kernel/perf_cpum_cf.c if (atomic_read(&num_events) == 0 && __kernel_cpumcf_begin()) atomic_read 452 arch/s390/kernel/perf_cpum_cf.c if (!atomic_read(&cpuhw->ctr_set[event->hw.config_base])) atomic_read 188 arch/s390/kernel/perf_cpum_cf_diag.c atomic_read(&cf_diag_events)); atomic_read 249 arch/s390/kernel/perf_cpum_cf_diag.c attr->sample_type, atomic_read(&cf_diag_events)); atomic_read 778 arch/s390/kernel/perf_cpum_sf.c if (atomic_read(&num_events) == 0 && reserve_pmc_hardware()) atomic_read 2087 arch/s390/kernel/perf_cpum_sf.c if (!atomic_read(&num_events)) atomic_read 372 arch/s390/kernel/time.c sw0 = atomic_read(sw_ptr); atomic_read 374 arch/s390/kernel/time.c sw1 = atomic_read(sw_ptr); atomic_read 422 arch/s390/kernel/time.c rc = (atomic_read(sw_ptr) & 0x80000000U) != 0; atomic_read 604 arch/s390/kernel/time.c while (atomic_read(&sync->cpus) != 0) atomic_read 361 arch/s390/kernel/topology.c if (atomic_read(&topology_poll) > 60) atomic_read 1814 arch/s390/kvm/interrupt.c online_vcpus = atomic_read(&kvm->online_vcpus); atomic_read 2987 arch/s390/kvm/interrupt.c int vcpu_id, online_vcpus = atomic_read(&kvm->online_vcpus); atomic_read 1919 arch/s390/kvm/kvm-s390.c int slot = atomic_read(&slots->lru_slot); atomic_read 2553 arch/s390/kvm/kvm-s390.c for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) atomic_read 3122 arch/s390/kvm/kvm-s390.c return atomic_read(&vcpu->arch.sie_block->prog20) & atomic_read 3732 arch/s390/kvm/kvm-s390.c cpuflags = atomic_read(&vcpu->arch.sie_block->cpuflags); atomic_read 4165 arch/s390/kvm/kvm-s390.c online_vcpus = atomic_read(&vcpu->kvm->online_vcpus); atomic_read 4205 arch/s390/kvm/kvm-s390.c online_vcpus = atomic_read(&vcpu->kvm->online_vcpus); atomic_read 60 arch/s390/kvm/kvm-s390.h return (atomic_read(&vcpu->arch.sie_block->cpuflags) & flags) == flags; atomic_read 811 arch/s390/kvm/priv.c cpus = atomic_read(&vcpu->kvm->online_vcpus); atomic_read 87 arch/s390/kvm/vsie.c return !(atomic_read(&vsie_page->scb_s.prog20) & PROG_REQUEST); atomic_read 96 arch/s390/kvm/vsie.c cpuflags = atomic_read(&vsie_page->scb_o->cpuflags); atomic_read 106 arch/s390/kvm/vsie.c int newflags, cpuflags = atomic_read(&scb_o->cpuflags); atomic_read 486 arch/s390/kvm/vsie.c if (!(atomic_read(&scb_s->cpuflags) & CPUSTAT_KSS)) atomic_read 491 arch/s390/kvm/vsie.c if (!(atomic_read(&scb_s->cpuflags) & CPUSTAT_SM)) atomic_read 1068 arch/s390/kvm/vsie.c if (!(atomic_read(&scb_o->cpuflags) & CPUSTAT_STOP_INT)) atomic_read 1236 arch/s390/kvm/vsie.c nr_vcpus = atomic_read(&kvm->online_vcpus); atomic_read 160 arch/s390/mm/pgalloc.c old = atomic_read(v); atomic_read 205 arch/s390/mm/pgalloc.c mask = atomic_read(&page->_refcount) >> 24; atomic_read 175 arch/sh/drivers/dma/dma-api.c if (atomic_read(&channel->busy) == 0) atomic_read 102 arch/sh/kernel/ftrace.c int old = atomic_read(&nmi_running); atomic_read 159 arch/sh/kernel/ftrace.c if (!atomic_read(&nmi_running)) atomic_read 164 arch/sh/kernel/ftrace.c } while (atomic_read(&nmi_running)); atomic_read 330 arch/sh/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 50 arch/sh/kernel/irq.c seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); atomic_read 138 arch/sh/kernel/perf_event.c if (atomic_read(&num_events) == 0 && atomic_read 364 arch/sh/kernel/smp.c if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) { atomic_read 396 arch/sh/kernel/smp.c if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) { atomic_read 439 arch/sh/kernel/smp.c if ((atomic_read(&vma->vm_mm->mm_users) != 1) || atomic_read 130 arch/sparc/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 128 arch/sparc/kernel/kgdb_64.c if (atomic_read(&kgdb_active) != -1) atomic_read 57 arch/sparc/kernel/nmi.c if (atomic_read(&nmi_active)) { atomic_read 167 arch/sparc/kernel/nmi.c if (!atomic_read(&nmi_active)) atomic_read 193 arch/sparc/kernel/nmi.c if (!atomic_read(&nmi_active)) { atomic_read 287 arch/sparc/kernel/nmi.c if (atomic_read(&nmi_active) == -1) { atomic_read 310 arch/sparc/kernel/nmi.c if (atomic_read(&nmi_active) == -1) atomic_read 1177 arch/sparc/kernel/perf_event.c if (atomic_read(&active_events) == 0) { atomic_read 1178 arch/sparc/kernel/perf_event.c if (atomic_read(&nmi_active) > 0) { atomic_read 1180 arch/sparc/kernel/perf_event.c BUG_ON(atomic_read(&nmi_active) != 0); atomic_read 1190 arch/sparc/kernel/perf_event.c if (atomic_read(&nmi_active) == 0) atomic_read 1425 arch/sparc/kernel/perf_event.c if (atomic_read(&nmi_active) < 0) atomic_read 1624 arch/sparc/kernel/perf_event.c if (!atomic_read(&active_events)) atomic_read 1085 arch/sparc/kernel/smp_64.c if (atomic_read(&mm->mm_users) == 1) { atomic_read 1123 arch/sparc/kernel/smp_64.c if (mm == current->mm && atomic_read(&mm->mm_users) == 1) atomic_read 1139 arch/sparc/kernel/smp_64.c if (mm == current->mm && atomic_read(&mm->mm_users) == 1) atomic_read 1184 arch/sparc/kernel/smp_64.c while (atomic_read(&smp_capture_registry) != ncpus) atomic_read 2025 arch/sparc/kernel/traps_64.c if ((cnt = atomic_read(ocnt)) != 0) { atomic_read 577 arch/sparc/mm/init_64.c atomic_read(&dcpage_flushes)); atomic_read 580 arch/sparc/mm/init_64.c atomic_read(&dcpage_flushes_xcall)); atomic_read 1693 arch/sparc/mm/srmmu.c if (atomic_read(&mm->mm_users) == 1 && current->active_mm == mm) atomic_read 61 arch/sparc/oprofile/init.c if (atomic_read(&nmi_active) <= 0) atomic_read 109 arch/um/drivers/port_kern.c if (atomic_read(&port->wait_count) == 0) { atomic_read 323 arch/um/kernel/process.c return atomic_read(&using_sysemu); atomic_read 528 arch/um/kernel/tlb.c if (atomic_read(¤t->mm->mm_users) == 0) atomic_read 556 arch/um/kernel/tlb.c if (atomic_read(&mm->mm_users) == 0) atomic_read 342 arch/x86/events/core.c if (atomic_read(&pmc_refcount) == 0) { atomic_read 383 arch/x86/events/core.c if (i != what && atomic_read(&x86_pmu.lbr_exclusive[i])) atomic_read 1562 arch/x86/events/core.c if (!atomic_read(&active_events)) atomic_read 2665 arch/x86/events/intel/core.c if (!atomic_read(&era->ref) || era->config == reg->config) { atomic_read 161 arch/x86/events/intel/uncore.c if (!atomic_read(&er->ref) || atomic_read 563 arch/x86/events/intel/uncore_nhmex.c if (!atomic_read(&er->ref) || er->config == config) { atomic_read 590 arch/x86/events/intel/uncore_nhmex.c if (__BITS_VALUE(atomic_read(&er->ref), idx, 8)) { atomic_read 597 arch/x86/events/intel/uncore_nhmex.c if (!atomic_read(&er->ref) || !((er->config ^ config) & mask)) { atomic_read 1004 arch/x86/events/intel/uncore_nhmex.c if (!atomic_read(&er->ref) || er->config == reg1->config) { atomic_read 1015 arch/x86/events/intel/uncore_nhmex.c if (!__BITS_VALUE(atomic_read(&er->ref), idx - 2, 8) || atomic_read 1023 arch/x86/events/intel/uncore_nhmex.c if (!atomic_read(&er->ref) || atomic_read 878 arch/x86/events/intel/uncore_snbep.c if (!__BITS_VALUE(atomic_read(&er->ref), i, 6) || atomic_read 1005 arch/x86/events/intel/uncore_snbep.c if (!__BITS_VALUE(atomic_read(&er->ref), idx, 8) || atomic_read 34 arch/x86/include/asm/mmu_context.h atomic_read(&mm->context.perf_rdpmc_allowed)) atomic_read 25 arch/x86/include/asm/qspinlock.h val |= atomic_read(&lock->val) & ~_Q_PENDING_MASK; atomic_read 98 arch/x86/include/asm/qspinlock.h while (atomic_read(&lock->val) != 0) atomic_read 101 arch/x86/include/asm/refcount.h c = atomic_read(&(r->refs)); atomic_read 852 arch/x86/include/asm/uv/uv_bau.h if (atomic_read(v) >= u) { atomic_read 406 arch/x86/kernel/apic/apic.c rsvd = atomic_read(&eilvt_offsets[offset]); atomic_read 629 arch/x86/kernel/cpu/mce/core.c if (atomic_read(&num_notifiers) > NUM_DEFAULT_NOTIFIERS) atomic_read 857 arch/x86/kernel/cpu/mce/core.c if (atomic_read(&mce_panicked)) atomic_read 978 arch/x86/kernel/cpu/mce/core.c while (atomic_read(&mce_callin) != cpus) { atomic_read 1004 arch/x86/kernel/cpu/mce/core.c while (atomic_read(&mce_executing) < order) { atomic_read 1017 arch/x86/kernel/cpu/mce/core.c *no_way_out = atomic_read(&global_nwo); atomic_read 1049 arch/x86/kernel/cpu/mce/core.c while (atomic_read(&mce_executing) <= cpus) { atomic_read 1063 arch/x86/kernel/cpu/mce/core.c while (atomic_read(&mce_executing) != 0) { atomic_read 200 arch/x86/kernel/cpu/mce/intel.c if (!atomic_read(&cmci_storm_on_cpus)) { atomic_read 289 arch/x86/kernel/cpu/mce/therm_throt.c if (!atomic_read(&therm_throt_en)) atomic_read 535 arch/x86/kernel/cpu/microcode/core.c while (atomic_read(t) < all_cpus) { atomic_read 538 arch/x86/kernel/cpu/microcode/core.c all_cpus - atomic_read(t)); atomic_read 2222 arch/x86/kernel/cpu/resctrl/rdtgroup.c if (atomic_read(&sentry->waitcount) != 0) atomic_read 2264 arch/x86/kernel/cpu/resctrl/rdtgroup.c if (atomic_read(&rdtgrp->waitcount) != 0) atomic_read 1042 arch/x86/kernel/ftrace.c if (unlikely(atomic_read(¤t->tracing_graph_pause))) atomic_read 163 arch/x86/kernel/irq.c seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); atomic_read 165 arch/x86/kernel/irq.c seq_printf(p, "%*s: %10u\n", prec, "MIS", atomic_read(&irq_mis_count)); atomic_read 223 arch/x86/kernel/irq.c u64 sum = atomic_read(&irq_err_count); atomic_read 502 arch/x86/kernel/kgdb.c if (atomic_read(&kgdb_active) != -1) { atomic_read 533 arch/x86/kernel/kgdb.c if (atomic_read(&kgdb_cpu_doing_single_step) != -1) { atomic_read 873 arch/x86/kernel/reboot.c while ((atomic_read(&waiting_for_crash_ipi) > 0) && msecs) { atomic_read 121 arch/x86/kernel/smp.c if (raw_smp_processor_id() == atomic_read(&stopping_cpu)) atomic_read 316 arch/x86/kernel/tboot.c while (atomic_read((atomic_t *)&tboot->num_in_wfs) != num_aps && atomic_read 325 arch/x86/kernel/tboot.c return !(atomic_read((atomic_t *)&tboot->num_in_wfs) == num_aps); atomic_read 332 arch/x86/kernel/tboot.c if (tboot_wait_for_aps(atomic_read(&ap_wfs_count))) atomic_read 576 arch/x86/kernel/traps.c if (unlikely(atomic_read(&modifying_ftrace_code)) && atomic_read 334 arch/x86/kernel/tsc_sync.c while (atomic_read(&start_count) != cpus - 1) { atomic_read 335 arch/x86/kernel/tsc_sync.c if (atomic_read(&skip_test) > 0) { atomic_read 349 arch/x86/kernel/tsc_sync.c while (atomic_read(&stop_count) != cpus-1) atomic_read 393 arch/x86/kernel/tsc_sync.c if (atomic_read(&test_runs) > 0) atomic_read 431 arch/x86/kernel/tsc_sync.c while (atomic_read(&start_count) != cpus) atomic_read 449 arch/x86/kernel/tsc_sync.c while (atomic_read(&stop_count) != cpus) atomic_read 462 arch/x86/kernel/tsc_sync.c if (!atomic_read(&test_runs)) atomic_read 182 arch/x86/kvm/hyperv.c gsi = atomic_read(&synic->sint_to_gsi[sint]); atomic_read 1338 arch/x86/kvm/hyperv.c if (likely(!atomic_read(&hv->num_mismatched_vp_indexes))) { atomic_read 247 arch/x86/kvm/i8254.c if (atomic_read(&ps->reinject) && !atomic_xchg(&ps->irq_ack, 0)) atomic_read 262 arch/x86/kvm/i8254.c if (atomic_read(&kvm->arch.vapics_in_nmi_mode) > 0) atomic_read 272 arch/x86/kvm/i8254.c if (atomic_read(&ps->reinject)) atomic_read 295 arch/x86/kvm/i8254.c if (atomic_read(&ps->reinject) == reinject) atomic_read 1552 arch/x86/kvm/lapic.c if (atomic_read(&apic->lapic_timer.pending)) atomic_read 1736 arch/x86/kvm/lapic.c if (atomic_read(&ktimer->pending)) { atomic_read 1756 arch/x86/kvm/lapic.c if (!apic_lvtt_period(apic) && atomic_read(&ktimer->pending)) atomic_read 1770 arch/x86/kvm/lapic.c if (!apic_lvtt_period(apic) && atomic_read(&apic->lapic_timer.pending)) atomic_read 2233 arch/x86/kvm/lapic.c return atomic_read(&apic->lapic_timer.pending); atomic_read 2357 arch/x86/kvm/lapic.c if (atomic_read(&apic->lapic_timer.pending) > 0) { atomic_read 5369 arch/x86/kvm/mmu.c return atomic_read(&sp->write_flooding_count) >= 3; atomic_read 1851 arch/x86/kvm/x86.c atomic_read(&vcpu->kvm->online_vcpus)); atomic_read 1866 arch/x86/kvm/x86.c atomic_read(&vcpu->kvm->online_vcpus), atomic_read 2229 arch/x86/kvm/x86.c atomic_read(&kvm->online_vcpus)); atomic_read 7230 arch/x86/kvm/x86.c atomic_read(&kvm_guest_has_master_clock) != 0) atomic_read 9157 arch/x86/kvm/x86.c if (kvm_check_tsc_unstable() && atomic_read(&kvm->online_vcpus) != 0) atomic_read 9608 arch/x86/kvm/x86.c for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) atomic_read 10278 arch/x86/kvm/x86.c return atomic_read(&kvm->arch.assigned_device_count); atomic_read 10296 arch/x86/kvm/x86.c return atomic_read(&kvm->arch.noncoherent_dma_count); atomic_read 78 arch/x86/mm/mmio-mod.c return atomic_read(&mmiotrace_enabled); atomic_read 479 arch/x86/platform/uv/uv_nmi.c nmi = atomic_read(&hub_nmi->in_nmi); atomic_read 507 arch/x86/platform/uv/uv_nmi.c nmi = atomic_read(&hub_nmi->in_nmi); atomic_read 517 arch/x86/platform/uv/uv_nmi.c nmi = atomic_read(&uv_in_nmi); atomic_read 539 arch/x86/platform/uv/uv_nmi.c if (cpu == atomic_read(&hub_nmi->cpu_owner)) { atomic_read 657 arch/x86/platform/uv/uv_nmi.c atomic_read(&uv_nmi_cpus_in_nmi), num_online_cpus()); atomic_read 725 arch/x86/platform/uv/uv_nmi.c while (atomic_read(&uv_nmi_cpus_in_nmi) > 0) atomic_read 729 arch/x86/platform/uv/uv_nmi.c while (atomic_read(&uv_nmi_slave_continue)) atomic_read 738 arch/x86/platform/uv/uv_nmi.c int in = atomic_read(&uv_nmi_cpus_in_nmi); atomic_read 744 arch/x86/platform/uv/uv_nmi.c while (!atomic_read(&uv_nmi_slave_continue)) atomic_read 760 arch/x86/platform/uv/uv_nmi.c atomic_read(&uv_nmi_cpus_in_nmi), cpu); atomic_read 778 arch/x86/platform/uv/uv_nmi.c while (!atomic_read(&uv_nmi_slave_continue)) atomic_read 815 arch/x86/platform/uv/uv_nmi.c while (atomic_read(&uv_nmi_kexec_failed) == 0) atomic_read 875 arch/x86/platform/uv/uv_nmi.c sig = atomic_read(&uv_nmi_slave_continue); atomic_read 911 arch/x86/platform/uv/uv_nmi.c master = (atomic_read(&uv_nmi_cpu) == cpu); atomic_read 50 arch/x86/xen/spinlock.c if (atomic_read(nest_cnt) == 1 && xen_test_irq_pending(irq)) { atomic_read 47 arch/xtensa/kernel/jump_label.c while (atomic_read(&patch->cpu_count) <= num_online_cpus()) atomic_read 5240 block/bfq-iosched.c if (atomic_read(&bic->icq.ioc->active_ref) == 0 || atomic_read 603 block/bio.c BIO_BUG_ON(!atomic_read(&bio->__bi_cnt)); atomic_read 1814 block/bio.c BUG_ON(atomic_read(&bio->__bi_remaining) <= 0); atomic_read 974 block/blk-cgroup.c if (blkcg_debug_stats && atomic_read(&blkg->use_delay)) { atomic_read 978 block/blk-cgroup.c atomic_read(&blkg->use_delay), atomic_read 1291 block/blk-cgroup.c if (ioc && atomic_read(&ioc->nr_tasks) > 1) atomic_read 1660 block/blk-cgroup.c int cur_use = atomic_read(&blkg->use_delay); atomic_read 1701 block/blk-cgroup.c if (atomic_read(&blkg->use_delay)) { atomic_read 965 block/blk-iocost.c ioc_gen = atomic_read(&ioc->hweight_gen); atomic_read 1092 block/blk-iocost.c iocg->hweight_gen = atomic_read(&ioc->hweight_gen) - 1; atomic_read 1237 block/blk-iocost.c if (!atomic_read(&blkg->use_delay) && atomic_read 1421 block/blk-iocost.c !atomic_read(&iocg_to_blkg(iocg)->use_delay) && atomic_read 99 block/blk-iolatency.c return atomic_read(&blkiolat->enabled) > 0; atomic_read 286 block/blk-iolatency.c unsigned use_delay = atomic_read(&lat_to_blkg(iolat)->use_delay); atomic_read 329 block/blk-iolatency.c unsigned long old = atomic_read(&lat_info->scale_cookie); atomic_read 396 block/blk-iolatency.c unsigned int our_cookie = atomic_read(&iolat->scale_cookie); atomic_read 409 block/blk-iolatency.c cur_cookie = atomic_read(&lat_info->scale_cookie); atomic_read 549 block/blk-iolatency.c atomic_read(&lat_info->scale_cookie) == DEFAULT_SCALE_COOKIE) atomic_read 683 block/blk-iolatency.c cookie = atomic_read(&lat_info->scale_cookie); atomic_read 992 block/blk-iolatency.c atomic_read(&parent->child_lat.scale_cookie)); atomic_read 99 block/blk-mq-debugfs.c seq_printf(m, "%d\n", atomic_read(&q->pm_only)); atomic_read 449 block/blk-mq-debugfs.c atomic_read(&tags->active_queues)); atomic_read 613 block/blk-mq-debugfs.c seq_printf(m, "%d\n", atomic_read(&hctx->nr_active)); atomic_read 87 block/blk-mq-tag.c users = atomic_read(&hctx->tags->active_queues); atomic_read 95 block/blk-mq-tag.c return atomic_read(&hctx->nr_active) < depth; atomic_read 11 block/blk-rq-qos.c unsigned int cur = atomic_read(v); atomic_read 753 block/blk-wbt.c atomic_read(&rwb->rq_wait[i].inflight)); atomic_read 82 block/blk-wbt.h ret += atomic_read(&rwb->rq_wait[i].inflight); atomic_read 686 drivers/acpi/apei/ghes.c count = atomic_read(&cache->count); atomic_read 1810 drivers/android/binder.c if (thread->is_dead && !atomic_read(&thread->tmp_ref)) { atomic_read 5639 drivers/android/binder.c atomic_read(&thread->tmp_ref)); atomic_read 5835 drivers/android/binder.c int temp = atomic_read(&stats->bc[i]); atomic_read 5845 drivers/android/binder.c int temp = atomic_read(&stats->br[i]); atomic_read 5857 drivers/android/binder.c int created = atomic_read(&stats->obj_created[i]); atomic_read 5858 drivers/android/binder.c int deleted = atomic_read(&stats->obj_deleted[i]); atomic_read 6048 drivers/android/binder.c unsigned int log_cur = atomic_read(&log->cur); atomic_read 713 drivers/atm/idt77252.c entries = atomic_read(&scq->used); atomic_read 768 drivers/atm/idt77252.c TXPRINTK("%d entries in SCQ used (push).\n", atomic_read(&scq->used)); atomic_read 771 drivers/atm/idt77252.c card->name, atomic_read(&scq->used), atomic_read 798 drivers/atm/idt77252.c card->name, atomic_read(&scq->used), scq->next); atomic_read 2545 drivers/atm/idt77252.c while (atomic_read(&vc->scq->used) > 0) { atomic_read 2549 drivers/atm/idt77252.c card->name, atomic_read(&vc->scq->used)); atomic_read 2835 drivers/atm/iphase.c printk("section_bip: %d\n", atomic_read(&stats->section_bip)); atomic_read 2836 drivers/atm/iphase.c printk("line_bip : %d\n", atomic_read(&stats->line_bip)); atomic_read 2837 drivers/atm/iphase.c printk("path_bip : %d\n", atomic_read(&stats->path_bip)); atomic_read 2838 drivers/atm/iphase.c printk("line_febe : %d\n", atomic_read(&stats->line_febe)); atomic_read 2839 drivers/atm/iphase.c printk("path_febe : %d\n", atomic_read(&stats->path_febe)); atomic_read 2840 drivers/atm/iphase.c printk("corr_hcs : %d\n", atomic_read(&stats->corr_hcs)); atomic_read 2841 drivers/atm/iphase.c printk("uncorr_hcs : %d\n", atomic_read(&stats->uncorr_hcs)); atomic_read 2842 drivers/atm/iphase.c printk("tx_cells : %d\n", atomic_read(&stats->tx_cells)); atomic_read 2843 drivers/atm/iphase.c printk("rx_cells : %d\n", atomic_read(&stats->rx_cells)); atomic_read 3060 drivers/atm/iphase.c if (atomic_read(&vcc->stats->tx) % 20 == 0) { atomic_read 54 drivers/atm/suni.c if (atomic_read(&stats->s) < 0) atomic_set(&stats->s,INT_MAX); atomic_read 166 drivers/atm/uPD98402.c if (atomic_read(&PRIV(dev)->sonet_stats.s) < 0) \ atomic_read 1142 drivers/auxdisplay/panel.c if (!atomic_read(&keypad_available)) { atomic_read 488 drivers/base/dd.c if (local_trigger_count != atomic_read(&deferred_trigger_count)) atomic_read 495 drivers/base/dd.c int local_trigger_count = atomic_read(&deferred_trigger_count); atomic_read 672 drivers/base/dd.c atomic_read(&probe_count)); atomic_read 673 drivers/base/dd.c if (atomic_read(&probe_count)) atomic_read 688 drivers/base/dd.c wait_event(probe_waitqueue, atomic_read(&probe_count) == 0); atomic_read 200 drivers/base/power/domain.c if (!WARN_ON(atomic_read(&genpd->sd_count) == 0)) atomic_read 516 drivers/base/power/domain.c atomic_read(&genpd->sd_count) > 0) atomic_read 550 drivers/base/power/domain.c if (atomic_read(&genpd->sd_count) > 0) atomic_read 967 drivers/base/power/domain.c || atomic_read(&genpd->sd_count) > 0) atomic_read 1352 drivers/base/power/main.c atomic_read(&dev->power.usage_count) > 1 || atomic_read 266 drivers/base/power/runtime.c else if (atomic_read(&dev->power.usage_count) > 0) atomic_read 269 drivers/base/power/runtime.c atomic_read(&dev->power.child_count)) atomic_read 1391 drivers/base/power/runtime.c atomic_read(&dev->power.child_count) > 0, atomic_read 1718 drivers/base/power/runtime.c return atomic_read(&dev->power.usage_count) <= 1 && atomic_read 1719 drivers/base/power/runtime.c (atomic_read(&dev->power.child_count) == 0 || atomic_read 510 drivers/base/power/sysfs.c return sprintf(buf, "%d\n", atomic_read(&dev->power.usage_count)); atomic_read 519 drivers/base/power/sysfs.c 0 : atomic_read(&dev->power.child_count)); atomic_read 51 drivers/base/power/wakeup.c unsigned int comb = atomic_read(&combined_event_count); atomic_read 872 drivers/base/power/wakeup.c return ret || atomic_read(&pm_abort_suspend) > 0; atomic_read 32 drivers/base/test/test_async_driver_probe.c if (atomic_read(&timeout)) { atomic_read 247 drivers/base/test/test_async_driver_probe.c if (atomic_read(&async_completed) != async_id) { atomic_read 251 drivers/base/test/test_async_driver_probe.c } else if (!atomic_read(&errors) && !atomic_read(&warnings)) { atomic_read 279 drivers/base/test/test_async_driver_probe.c atomic_read(&errors), atomic_read(&warnings)); atomic_read 181 drivers/block/aoe/aoecmd.c if (skb && atomic_read(&skb_shinfo(skb)->dataref) == 1) { atomic_read 235 drivers/block/aoe/aoecmd.c if (atomic_read(&skb_shinfo(skb)->dataref) != 1) { atomic_read 412 drivers/block/aoe/aoedev.c while (atomic_read(&skb_shinfo(skb)->dataref) != 1 && i-- > 0) atomic_read 181 drivers/block/drbd/drbd_actlog.c D_ASSERT(device, atomic_read(&device->md_io.in_use) == 1); atomic_read 249 drivers/block/drbd/drbd_actlog.c D_ASSERT(device, atomic_read(&device->local_cnt) > 0); atomic_read 268 drivers/block/drbd/drbd_actlog.c D_ASSERT(device, atomic_read(&device->local_cnt) > 0); atomic_read 657 drivers/block/drbd/drbd_actlog.c D_ASSERT(device, atomic_read(&device->local_cnt)); atomic_read 1164 drivers/block/drbd/drbd_bitmap.c if (atomic_read(&ctx->in_flight)) atomic_read 142 drivers/block/drbd/drbd_debugfs.c if (atomic_read(&tmp.in_use)) { atomic_read 166 drivers/block/drbd/drbd_debugfs.c int n = atomic_read(&device->ap_actlog_cnt); atomic_read 203 drivers/block/drbd/drbd_debugfs.c in_flight = atomic_read(&ctx->in_flight); atomic_read 2029 drivers/block/drbd/drbd_int.h if (atomic_read(&device->which) < 0) \ atomic_read 2032 drivers/block/drbd/drbd_int.h atomic_read(&device->which)) atomic_read 2262 drivers/block/drbd/drbd_int.h if (atomic_read(&device->suspend_cnt)) atomic_read 2275 drivers/block/drbd/drbd_int.h if (atomic_read(&device->ap_bio_cnt) > mxb) atomic_read 2299 drivers/block/drbd/drbd_main.c expect(atomic_read(&req->completion_ref) == 0) && atomic_read 2306 drivers/block/drbd/drbd_main.c req, atomic_read(&req->completion_ref), atomic_read 2746 drivers/block/drbd/drbd_main.c if (atomic_read(&connection->current_epoch->epoch_size) != 0) atomic_read 2747 drivers/block/drbd/drbd_main.c drbd_err(connection, "epoch_size:%d\n", atomic_read(&connection->current_epoch->epoch_size)); atomic_read 3574 drivers/block/drbd/drbd_main.c int cnt = atomic_read(&device->ap_bio_cnt); atomic_read 3638 drivers/block/drbd/drbd_main.c if (flags == BM_LOCKED_CHANGE_ALLOWED || atomic_read(&device->ap_bio_cnt) == 0) { atomic_read 715 drivers/block/drbd/drbd_nl.c wait_event(device->misc_wait, atomic_read(&device->ap_pending_cnt) == 0); atomic_read 911 drivers/block/drbd/drbd_nl.c wait_event(device->misc_wait, !atomic_read(&device->ap_bio_cnt)); atomic_read 1559 drivers/block/drbd/drbd_nl.c if (atomic_read(&device->ap_bio_cnt)) atomic_read 1969 drivers/block/drbd/drbd_nl.c wait_event(device->misc_wait, !atomic_read(&device->ap_pending_cnt) || drbd_suspended(device)); atomic_read 3399 drivers/block/drbd/drbd_nl.c s->dev_upper_pending = atomic_read(&device->ap_bio_cnt); atomic_read 3400 drivers/block/drbd/drbd_nl.c s->dev_lower_pending = atomic_read(&device->local_cnt); atomic_read 3644 drivers/block/drbd/drbd_nl.c s->peer_dev_pending = atomic_read(&device->ap_pending_cnt) + atomic_read 3645 drivers/block/drbd/drbd_nl.c atomic_read(&device->rs_pending_cnt); atomic_read 3646 drivers/block/drbd/drbd_nl.c s->peer_dev_unacked = atomic_read(&device->unacked_cnt); atomic_read 3842 drivers/block/drbd/drbd_nl.c nla_put_u32(skb, T_ap_bio_cnt, atomic_read(&device->ap_bio_cnt)) || atomic_read 3843 drivers/block/drbd/drbd_nl.c nla_put_u32(skb, T_ap_pending_cnt, atomic_read(&device->ap_pending_cnt)) || atomic_read 3844 drivers/block/drbd/drbd_nl.c nla_put_u32(skb, T_rs_pending_cnt, atomic_read(&device->rs_pending_cnt))) atomic_read 294 drivers/block/drbd/drbd_proc.c atomic_read(&device->local_cnt), atomic_read 295 drivers/block/drbd/drbd_proc.c atomic_read(&device->ap_pending_cnt) + atomic_read 296 drivers/block/drbd/drbd_proc.c atomic_read(&device->rs_pending_cnt), atomic_read 297 drivers/block/drbd/drbd_proc.c atomic_read(&device->unacked_cnt), atomic_read 298 drivers/block/drbd/drbd_proc.c atomic_read(&device->ap_bio_cnt), atomic_read 319 drivers/block/drbd/drbd_proc.c seq_printf(seq, "\tblocked on activity log: %d\n", atomic_read(&device->ap_actlog_cnt)); atomic_read 228 drivers/block/drbd/drbd_receiver.c if (!atomic_read(&device->pp_in_use_by_net)) atomic_read 274 drivers/block/drbd/drbd_receiver.c if (atomic_read(&device->pp_in_use) < mxb) atomic_read 279 drivers/block/drbd/drbd_receiver.c if (page && atomic_read(&device->pp_in_use_by_net) > 512) atomic_read 287 drivers/block/drbd/drbd_receiver.c if (atomic_read(&device->pp_in_use) < mxb) { atomic_read 411 drivers/block/drbd/drbd_receiver.c D_ASSERT(device, atomic_read(&peer_req->pending_bios) == 0); atomic_read 1368 drivers/block/drbd/drbd_receiver.c epoch_size = atomic_read(&epoch->epoch_size); atomic_read 1383 drivers/block/drbd/drbd_receiver.c atomic_read(&epoch->active) == 0 && atomic_read 1806 drivers/block/drbd/drbd_receiver.c if (atomic_read(&connection->current_epoch->epoch_size)) { atomic_read 1824 drivers/block/drbd/drbd_receiver.c if (atomic_read(&connection->current_epoch->epoch_size)) { atomic_read 2805 drivers/block/drbd/drbd_receiver.c atomic_read(&device->rs_sect_ev); atomic_read 2807 drivers/block/drbd/drbd_receiver.c if (atomic_read(&device->ap_actlog_cnt) atomic_read 5281 drivers/block/drbd/drbd_receiver.c i = atomic_read(&device->pp_in_use_by_net); atomic_read 5284 drivers/block/drbd/drbd_receiver.c i = atomic_read(&device->pp_in_use); atomic_read 5890 drivers/block/drbd/drbd_receiver.c atomic_read(&device->ap_in_flight) == 0 && atomic_read 98 drivers/block/drbd/drbd_req.c atomic_read(&req->completion_ref) || atomic_read 102 drivers/block/drbd/drbd_req.c s, atomic_read(&req->completion_ref)); atomic_read 262 drivers/block/drbd/drbd_req.c req->epoch == atomic_read(&first_peer_device(device)->connection->current_tle_nr)) atomic_read 930 drivers/block/drbd/drbd_req.c return atomic_read(&device->local_cnt) > atomic_read 931 drivers/block/drbd/drbd_req.c atomic_read(&device->ap_pending_cnt) + atomic_read(&device->rs_pending_cnt); atomic_read 1016 drivers/block/drbd/drbd_req.c atomic_read(&device->ap_in_flight) >= nc->cong_fill) { atomic_read 1364 drivers/block/drbd/drbd_req.c req->epoch = atomic_read(&first_peer_device(device)->connection->current_tle_nr); atomic_read 1701 drivers/block/drbd/drbd_worker.c if (atomic_read(&device->unacked_cnt) || atomic_read(&device->rs_pending_cnt)) { atomic_read 2125 drivers/block/drbd/drbd_worker.c atomic_read(&connection->current_tle_nr) != atomic_read 880 drivers/block/floppy.c if (WARN(atomic_read(&usage_count) == 0, atomic_read 2899 drivers/block/floppy.c if (WARN(atomic_read(&usage_count) == 0, atomic_read 4202 drivers/block/floppy.c if (WARN(atomic_read(&usage_count) == 0, atomic_read 4726 drivers/block/floppy.c if (atomic_read(&usage_count)) atomic_read 4986 drivers/block/floppy.c if (atomic_read(&usage_count)) atomic_read 1246 drivers/block/loop.c if (atomic_read(&lo->lo_refcnt) > 1) { atomic_read 2202 drivers/block/loop.c if (atomic_read(&lo->lo_refcnt) > 0) { atomic_read 762 drivers/block/mtip32xx/mtip32xx.c WARN_ON_ONCE(atomic_read(&dd->irq_workers_active) != 0); atomic_read 2675 drivers/block/mtip32xx/mtip32xx.c } while (atomic_read(&dd->irq_workers_active) != 0 && atomic_read 2678 drivers/block/mtip32xx/mtip32xx.c if (atomic_read(&dd->irq_workers_active) != 0) atomic_read 4205 drivers/block/mtip32xx/mtip32xx.c } while (atomic_read(&dd->irq_workers_active) != 0 && atomic_read 4211 drivers/block/mtip32xx/mtip32xx.c if (atomic_read(&dd->irq_workers_active) != 0) { atomic_read 401 drivers/block/nbd.c atomic_read(&config->live_connections), atomic_read 867 drivers/block/nbd.c atomic_read(&config->live_connections) > 0, atomic_read 1308 drivers/block/nbd.c atomic_read(&config->recv_threads) == 0); atomic_read 747 drivers/block/paride/pt.c if (atomic_read(&tape->available) > 1) atomic_read 511 drivers/block/pktcdvd.c BUG_ON(atomic_read(&pd->cdrw.pending_bios) <= 0); atomic_read 844 drivers/block/pktcdvd.c if (atomic_read(&pd->iosched.attention) == 0) atomic_read 869 drivers/block/pktcdvd.c if (atomic_read(&pd->cdrw.pending_bios) > 0) { atomic_read 878 drivers/block/pktcdvd.c if (atomic_read(&pd->cdrw.pending_bios) > 0) { atomic_read 1338 drivers/block/pktcdvd.c if (atomic_read(&pkt->io_wait) > 0) atomic_read 1341 drivers/block/pktcdvd.c if (atomic_read(&pkt->io_errors) > 0) { atomic_read 1349 drivers/block/pktcdvd.c if (atomic_read(&pkt->io_wait) > 0) atomic_read 1387 drivers/block/pktcdvd.c if (atomic_read(&pkt->run_sm) > 0) { atomic_read 1447 drivers/block/pktcdvd.c if (atomic_read(&pd->scan_queue) > 0) atomic_read 1452 drivers/block/pktcdvd.c if (atomic_read(&pkt->run_sm) > 0) atomic_read 1457 drivers/block/pktcdvd.c if (atomic_read(&pd->iosched.attention) != 0) atomic_read 2551 drivers/block/pktcdvd.c seq_printf(m, "\tbios pending:\t\t%d\n", atomic_read(&pd->cdrw.pending_bios)); atomic_read 147 drivers/block/rsxx/core.c i, atomic_read(&card->ctrl[i].stats.hw_q_depth)); atomic_read 126 drivers/block/rsxx/dev.c if (atomic_read(&meta->error)) atomic_read 201 drivers/block/rsxx/dma.c q_depth += atomic_read(&card->ctrl[i].stats.hw_q_depth); atomic_read 348 drivers/block/rsxx/dma.c if (atomic_read(&ctrl->stats.hw_q_depth) == 0 || atomic_read 565 drivers/block/rsxx/dma.c if (atomic_read(&ctrl->stats.hw_q_depth) == 0) atomic_read 746 drivers/block/xen-blkback/blkback.c if (atomic_dec_and_test(&ring->inflight) && atomic_read(&blkif->drain)) { atomic_read 1067 drivers/block/xen-blkback/blkback.c if (atomic_read(&ring->inflight) == 0) atomic_read 1072 drivers/block/xen-blkback/blkback.c if (!atomic_read(&blkif->drain)) atomic_read 267 drivers/block/xen-blkback/xenbus.c if (atomic_read(&ring->inflight) > 0) { atomic_read 299 drivers/block/xen-blkback/xenbus.c BUG_ON(atomic_read(&ring->persistent_gnt_in_use) != 0); atomic_read 146 drivers/bluetooth/bcm203x.c if (atomic_read(&data->shutdown)) atomic_read 160 drivers/bluetooth/bfusb.c while ((atomic_read(&data->pending_tx) < BFUSB_MAX_BULK_TX) && atomic_read 76 drivers/char/agp/backend.c if (atomic_read(&bridge->agp_in_use)) atomic_read 752 drivers/char/agp/frontend.c if (atomic_read(&agp_bridge->agp_in_use)) atomic_read 229 drivers/char/agp/generic.c cur_memory = atomic_read(&bridge->current_memory_agp); atomic_read 382 drivers/char/agp/generic.c info->current_memory = atomic_read(&bridge->current_memory_agp); atomic_read 518 drivers/char/apm-emulation.c if (atomic_read(&userspace_notification_inhibit)) atomic_read 556 drivers/char/apm-emulation.c atomic_read(&suspend_acks_pending) == 0, atomic_read 165 drivers/char/ipmi/ipmb_dev_int.c if (atomic_read(&ipmb_dev->request_queue_len)) atomic_read 184 drivers/char/ipmi/ipmb_dev_int.c if (atomic_read(&ipmb_dev->request_queue_len) >= atomic_read 651 drivers/char/ipmi/ipmi_msghandler.c ((unsigned int) atomic_read(&(intf)->stats[IPMI_STAT_ ## stat])) atomic_read 4764 drivers/char/ipmi/ipmi_msghandler.c if (atomic_read(&stop_operation)) atomic_read 4769 drivers/char/ipmi/ipmi_msghandler.c if (atomic_read(&intf->event_waiters)) { atomic_read 4885 drivers/char/ipmi/ipmi_msghandler.c while (atomic_read(&panic_done_count) != 0) atomic_read 5203 drivers/char/ipmi/ipmi_msghandler.c count = atomic_read(&smi_msg_inuse_count); atomic_read 5206 drivers/char/ipmi/ipmi_msghandler.c count = atomic_read(&recv_msg_inuse_count); atomic_read 159 drivers/char/ipmi/ipmi_poweroff.c while (atomic_read(&dummy_count) > 0) { atomic_read 250 drivers/char/ipmi/ipmi_si_intf.c ((unsigned int) atomic_read(&(smi)->stats[SI_STAT_ ## stat])) atomic_read 833 drivers/char/ipmi/ipmi_si_intf.c && (atomic_read(&smi_info->req_events))) { atomic_read 1028 drivers/char/ipmi/ipmi_si_intf.c if (atomic_read(&smi_info->need_watch)) { atomic_read 302 drivers/char/ipmi/ipmi_ssif.c ((unsigned int) atomic_read(&(ssif)->stats[SSIF_STAT_ ## stat])) atomic_read 525 drivers/char/ipmi/ipmi_watchdog.c while (atomic_read(&panic_done_count) != 0) atomic_read 538 drivers/char/ipmi/ipmi_watchdog.c while (atomic_read(&panic_done_count) != 0) atomic_read 1080 drivers/char/ipmi/ipmi_watchdog.c while (atomic_read(&msg_tofree)) atomic_read 612 drivers/char/ppdev.c ret = atomic_read(&pp->irqc); atomic_read 782 drivers/char/ppdev.c if (atomic_read(&pp->irqc)) atomic_read 615 drivers/clocksource/arm_arch_timer.c return atomic_read(&timer_unstable_counter_workaround_in_use); atomic_read 69 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 98 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 123 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 160 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 184 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 216 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 242 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 275 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 317 drivers/connector/cn_proc.c if (atomic_read(&proc_event_num_listeners) < 1) atomic_read 139 drivers/connector/cn_queue.c while (atomic_read(&dev->refcnt)) { atomic_read 141 drivers/connector/cn_queue.c dev->name, atomic_read(&dev->refcnt)); atomic_read 155 drivers/cpuidle/coupled.c while (atomic_read(a) < n) atomic_read 163 drivers/cpuidle/coupled.c while (atomic_read(a) > n) atomic_read 245 drivers/cpuidle/coupled.c int r = atomic_read(&coupled->ready_waiting_counts) >> WAITING_BITS; atomic_read 257 drivers/cpuidle/coupled.c int r = atomic_read(&coupled->ready_waiting_counts) >> WAITING_BITS; atomic_read 269 drivers/cpuidle/coupled.c int w = atomic_read(&coupled->ready_waiting_counts) & WAITING_MASK; atomic_read 281 drivers/cpuidle/coupled.c int w = atomic_read(&coupled->ready_waiting_counts) & WAITING_MASK; atomic_read 234 drivers/crypto/atmel-ecc.c tfm_cnt = atomic_read(&i2c_priv->tfm_count); atomic_read 363 drivers/crypto/atmel-ecc.c if (atomic_read(&i2c_priv->tfm_count)) { atomic_read 123 drivers/crypto/atmel-sha204a.c if (atomic_read(&i2c_priv->tfm_count)) { atomic_read 374 drivers/crypto/bcm/util.c atomic_read(&ipriv->session_count)); atomic_read 377 drivers/crypto/bcm/util.c atomic_read(&ipriv->stream_count)); atomic_read 380 drivers/crypto/bcm/util.c atomic_read(&ipriv->setkey_cnt[SPU_OP_CIPHER])); atomic_read 383 drivers/crypto/bcm/util.c atomic_read(&ipriv->op_counts[SPU_OP_CIPHER])); atomic_read 386 drivers/crypto/bcm/util.c op_cnt = atomic_read(&ipriv->cipher_cnt[alg][mode]); atomic_read 397 drivers/crypto/bcm/util.c atomic_read(&ipriv->op_counts[SPU_OP_HASH])); atomic_read 399 drivers/crypto/bcm/util.c op_cnt = atomic_read(&ipriv->hash_cnt[alg]); atomic_read 409 drivers/crypto/bcm/util.c atomic_read(&ipriv->setkey_cnt[SPU_OP_HMAC])); atomic_read 412 drivers/crypto/bcm/util.c atomic_read(&ipriv->op_counts[SPU_OP_HMAC])); atomic_read 414 drivers/crypto/bcm/util.c op_cnt = atomic_read(&ipriv->hmac_cnt[alg]); atomic_read 424 drivers/crypto/bcm/util.c atomic_read(&ipriv->setkey_cnt[SPU_OP_AEAD])); atomic_read 428 drivers/crypto/bcm/util.c atomic_read(&ipriv->op_counts[SPU_OP_AEAD])); atomic_read 430 drivers/crypto/bcm/util.c op_cnt = atomic_read(&ipriv->aead_cnt[alg]); atomic_read 446 drivers/crypto/bcm/util.c atomic_read(&ipriv->mb_no_spc)); atomic_read 449 drivers/crypto/bcm/util.c atomic_read(&ipriv->mb_send_fail)); atomic_read 452 drivers/crypto/bcm/util.c atomic_read(&ipriv->bad_icv)); atomic_read 151 drivers/crypto/caam/caamrng.c if (atomic_read(&bd->empty)) { atomic_read 153 drivers/crypto/caam/caamrng.c if (atomic_read(&bd->empty) == BUF_EMPTY) { atomic_read 164 drivers/crypto/caam/caamrng.c if (atomic_read(&bd->empty)) atomic_read 254 drivers/crypto/caam/caamrng.c if (atomic_read(&bd->empty) == BUF_PENDING) atomic_read 132 drivers/crypto/caam/jr.c if (atomic_read(&jrpriv->tfm_count)) { atomic_read 294 drivers/crypto/caam/jr.c tfm_cnt = atomic_read(&jrpriv->tfm_count); atomic_read 279 drivers/crypto/cavium/nitrox/nitrox_dev.h return atomic_read(&ndev->state) == __NDEV_READY; atomic_read 284 drivers/crypto/cavium/nitrox/nitrox_dev.h return atomic_read(&vfdev->state) == __NDEV_READY; atomic_read 321 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c if (!atomic_read(&cmdq->backlog_count)) atomic_read 542 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c budget = atomic_read(&cmdq->pending_count); atomic_read 549 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c if (atomic_read(&sr->status) != REQ_POSTED) atomic_read 601 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c if (atomic_read(&cmdq->backlog_count)) atomic_read 59 drivers/crypto/chelsio/chcr_core.c if (atomic_read(&dev->inflight)) { atomic_read 63 drivers/crypto/chelsio/chcr_core.c atomic_read(&dev->inflight)); atomic_read 68 drivers/crypto/chelsio/chcr_core.c atomic_read(&dev->inflight)); atomic_read 249 drivers/crypto/chelsio/chcr_core.c if (atomic_read(&dev->inflight) != 0) { atomic_read 1091 drivers/crypto/hisilicon/qm.c if (unlikely(atomic_read(&qp->qp_status.used) == QM_Q_DEPTH)) atomic_read 1332 drivers/crypto/hisilicon/qm.c while (atomic_read(&qp->qp_status.used)) { atomic_read 846 drivers/crypto/hisilicon/sec/sec_drv.c return !atomic_read(&msg_ring->used); atomic_read 866 drivers/crypto/hisilicon/sec/sec_drv.c if (write == read && atomic_read(&msg_ring->used) == SEC_QUEUE_LEN) { atomic_read 888 drivers/crypto/hisilicon/sec/sec_drv.c return SEC_QUEUE_LEN - atomic_read(&msg_ring->used) >= num; atomic_read 888 drivers/crypto/ixp4xx_crypto.c if (atomic_read(&ctx->configuring)) atomic_read 996 drivers/crypto/ixp4xx_crypto.c if (atomic_read(&ctx->configuring)) atomic_read 736 drivers/crypto/marvell/cesa.h u32 load = atomic_read(&engine->load); atomic_read 425 drivers/crypto/qat/qat_common/adf_dev_mgr.c return atomic_read(&accel_dev->ref_count) != 0; atomic_read 74 drivers/crypto/qat/qat_common/qat_crypto.c for (i = 0; i < atomic_read(&inst->refctr); i++) atomic_read 108 drivers/crypto/qat/qat_common/qat_crypto.c ctr = atomic_read(&tmp_dev->ref_count); atomic_read 135 drivers/crypto/qat/qat_common/qat_crypto.c ctr = atomic_read(&tmp_inst->refctr); atomic_read 188 drivers/crypto/ux500/cryp/cryp_core.c ctx->session_id = atomic_read(&session_id); atomic_read 407 drivers/crypto/ux500/cryp/cryp_core.c ctx->session_id != atomic_read(&session_id)) { atomic_read 117 drivers/crypto/virtio/virtio_crypto_mgr.c return atomic_read(&vcrypto_dev->ref_count) != 0; atomic_read 198 drivers/crypto/virtio/virtio_crypto_mgr.c ctr = atomic_read(&tmp_dev->ref_count); atomic_read 107 drivers/dma-buf/dma-fence-array.c return atomic_read(&array->num_pending) <= 0; atomic_read 1085 drivers/dma/dmaengine.c if (atomic_read(idr_ref) == 0) { atomic_read 238 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c atomic_read(&chan->descs_allocated)); atomic_read 370 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c axi_chan_name(chan), atomic_read(&chan->descs_allocated)); atomic_read 429 drivers/dma/mediatek/mtk-hsdma.c reserved = min_t(u16, num_sgs, atomic_read(&pc->nr_free)); atomic_read 629 drivers/dma/mediatek/mtk-hsdma.c if (atomic_read(&pc->nr_free) >= MTK_DMA_SIZE - 1) atomic_read 30 drivers/dma/qcom/hidma_dbg.c seq_printf(s, "allocated=%d\n", atomic_read(&tre->allocated)); atomic_read 70 drivers/dma/qcom/hidma_dbg.c atomic_read(&lldev->pending_tre_count)); atomic_read 126 drivers/dma/qcom/hidma_ll.c if (atomic_read(&tre->allocated) != true) { atomic_read 306 drivers/dma/qcom/hidma_ll.c while (atomic_read(&lldev->pending_tre_count)) { atomic_read 612 drivers/dma/qcom/hidma_ll.c if (atomic_read(&tre->allocated) != true) { atomic_read 570 drivers/dma/sun6i-dma.c if (!atomic_read(&sdev->tasklet_shutdown)) atomic_read 61 drivers/edac/edac_pci_sysfs.c return sprintf(data, "%u\n", atomic_read(&pci->counters.pe_count)); atomic_read 67 drivers/edac/edac_pci_sysfs.c return sprintf(data, "%u\n", atomic_read(&pci->counters.npe_count)); atomic_read 649 drivers/edac/edac_pci_sysfs.c before_count = atomic_read(&pci_parity_count); atomic_read 661 drivers/edac/edac_pci_sysfs.c if (before_count != atomic_read(&pci_parity_count)) atomic_read 430 drivers/edac/thunderx_edac.c while (!atomic_read(&lmc->ecc_int) && timeout--) { atomic_read 314 drivers/firewire/core-card.c atomic_read(&root_device->state) == FW_DEVICE_RUNNING; atomic_read 1001 drivers/firewire/core-device.c atomic_read(&device->state) == FW_DEVICE_INITIALIZING) { atomic_read 1179 drivers/firewire/core-device.c atomic_read(&device->state) == FW_DEVICE_INITIALIZING) { atomic_read 1283 drivers/firewire/core-device.c if (atomic_read(&device->state) == FW_DEVICE_RUNNING) { atomic_read 144 drivers/firewire/nosy.c atomic_read(&buffer->size) > 0) || atomic_read 148 drivers/firewire/nosy.c if (atomic_read(&buffer->size) == 0) atomic_read 188 drivers/firewire/nosy.c atomic_read(&buffer->size) + sizeof(struct packet) + length) { atomic_read 326 drivers/firewire/nosy.c if (atomic_read(&client->buffer.size) > 0) atomic_read 612 drivers/firmware/efi/efi.c for (i = 0; i < atomic_read(&rsv->count); i++) { atomic_read 1101 drivers/gpio/gpio-pca953x.c if (atomic_read(&chip->wakeup_path)) atomic_read 1114 drivers/gpio/gpio-pca953x.c if (!atomic_read(&chip->wakeup_path)) { atomic_read 550 drivers/gpio/gpio-rcar.c if (atomic_read(&p->wakeup_path)) atomic_read 1331 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c is_invalid_userptr = atomic_read(&mem->invalid); atomic_read 1583 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (atomic_read(&adev->gmc.vm_fault_info_updated) == 1) { atomic_read 1698 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (!atomic_read(&mem->invalid)) atomic_read 1724 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c invalid = atomic_read(&mem->invalid); atomic_read 1886 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c evicted_bos = atomic_read(&process_info->evicted_bos); atomic_read 1909 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (atomic_read(&process_info->evicted_bos) != evicted_bos) atomic_read 132 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (atomic_read(&p->ctx->guilty) == 1) { atomic_read 117 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c ctx->reset_counter = atomic_read(&adev->gpu_reset_counter); atomic_read 119 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c ctx->vram_lost_counter = atomic_read(&adev->vram_lost_counter); atomic_read 336 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c reset_counter = atomic_read(&adev->gpu_reset_counter); atomic_read 370 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c if (ctx->reset_counter != atomic_read(&adev->gpu_reset_counter)) atomic_read 373 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c if (ctx->vram_lost_counter != atomic_read(&adev->vram_lost_counter)) atomic_read 376 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c if (atomic_read(&ctx->guilty)) atomic_read 938 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c last_seq = atomic_read(&ring->fence_drv.last_seq); atomic_read 1003 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c if (preempt_seq <= atomic_read(&drv->last_seq)) atomic_read 1061 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c if (atomic_read(&ring->fence_drv.last_seq) != atomic_read 3938 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c dev_info(tmp_adev->dev, "GPU reset(%d) failed\n", atomic_read(&adev->gpu_reset_counter)); atomic_read 3941 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c dev_info(tmp_adev->dev, "GPU reset(%d) succeeded!\n", atomic_read(&adev->gpu_reset_counter)); atomic_read 122 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c seq = atomic_read(&drv->last_seq); atomic_read 241 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c last_seq = atomic_read(&ring->fence_drv.last_seq); atomic_read 372 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c emitted -= atomic_read(&ring->fence_drv.last_seq); atomic_read 406 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c amdgpu_fence_write(ring, atomic_read(&ring->fence_drv.last_seq)); atomic_read 713 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c atomic_read(&ring->fence_drv.last_seq)); atomic_read 182 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c atomic_read(&adev->gpu_reset_counter); atomic_read 565 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c return !!atomic_read(&src->enabled_types[type]); atomic_read 47 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), atomic_read 83 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c (*job)->vram_lost_counter = atomic_read(&adev->vram_lost_counter); atomic_read 230 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c if (job->vram_lost_counter != atomic_read(&ring->adev->vram_lost_counter)) atomic_read 918 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c ui32 = atomic_read(&adev->vram_lost_counter); atomic_read 187 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c || atomic_read(&ring->num_jobs[i])) { atomic_read 355 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c if (atomic_read(&adev->uvd.handles[i])) atomic_read 425 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c uint32_t handle = atomic_read(&adev->uvd.handles[i]); atomic_read 747 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c if (atomic_read(&adev->uvd.handles[i]) == handle) { atomic_read 771 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c if (atomic_read(&adev->uvd.handles[i]) == handle) { atomic_read 1287 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c if (atomic_read(&adev->uvd.handles[i])) atomic_read 267 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c if (atomic_read(&adev->vce.handles[i])) atomic_read 407 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t handle = atomic_read(&adev->vce.handles[i]); atomic_read 680 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c if (atomic_read(&p->adev->vce.handles[i]) == handle) { atomic_read 1088 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c atomic_read(&adev->gpu_reset_counter); atomic_read 1783 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c enable = !!atomic_read(&adev->vm_manager.num_prt_users); atomic_read 1275 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c && !atomic_read(&adev->gmc.vm_fault_info_updated)) { atomic_read 1460 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c && !atomic_read(&adev->gmc.vm_fault_info_updated)) { atomic_read 788 drivers/gpu/drm/amd/amdkfd/kfd_device.c return (atomic_read(&kfd_locked) > 0); atomic_read 1019 drivers/gpu/drm/amd/amdkfd/kfd_events.c int reset_cause = atomic_read(&dev->sram_ecc_flag) ? atomic_read 202 drivers/gpu/drm/arm/hdlcd_drv.c seq_printf(m, "underrun : %d\n", atomic_read(&hdlcd->buffer_underrun_count)); atomic_read 203 drivers/gpu/drm/arm/hdlcd_drv.c seq_printf(m, "dma_end : %d\n", atomic_read(&hdlcd->dma_end_count)); atomic_read 204 drivers/gpu/drm/arm/hdlcd_drv.c seq_printf(m, "bus_error: %d\n", atomic_read(&hdlcd->bus_error_count)); atomic_read 205 drivers/gpu/drm/arm/hdlcd_drv.c seq_printf(m, "vsync : %d\n", atomic_read(&hdlcd->vsync_count)); atomic_read 183 drivers/gpu/drm/arm/malidp_drv.c atomic_read(&malidp->config_valid) == MALIDP_CONFIG_VALID_DONE, atomic_read 1328 drivers/gpu/drm/arm/malidp_hw.c if ((atomic_read(&malidp->config_valid) != MALIDP_CONFIG_START) || atomic_read 1311 drivers/gpu/drm/drm_bufs.c if (atomic_read(&dev->buf_alloc)) { atomic_read 1503 drivers/gpu/drm/drm_bufs.c if (atomic_read(&dev->buf_alloc)) { atomic_read 294 drivers/gpu/drm/drm_dp_aux_dev.c wait_var_event(&aux_dev->usecount, !atomic_read(&aux_dev->usecount)); atomic_read 403 drivers/gpu/drm/drm_vblank.c if (atomic_read(&vblank->refcount) == 0 && vblank->enabled) { atomic_read 1043 drivers/gpu/drm/drm_vblank.c if (WARN_ON(atomic_read(&vblank->refcount) == 0)) atomic_read 1291 drivers/gpu/drm/drm_vblank.c if (atomic_read(&vblank->refcount) != 0 || drm_vblank_offdelay == 0) atomic_read 1787 drivers/gpu/drm/drm_vblank.c !atomic_read(&vblank->refcount)); atomic_read 73 drivers/gpu/drm/etnaviv/etnaviv_gem.h return atomic_read(&etnaviv_obj->gpu_active) != 0; atomic_read 1804 drivers/gpu/drm/etnaviv/etnaviv_gpu.c if (atomic_read(&gpu->sched.hw_rq_count)) atomic_read 98 drivers/gpu/drm/exynos/exynos7_drm_decon.c !atomic_read(&ctx->wait_vsync_event), atomic_read 602 drivers/gpu/drm/exynos/exynos7_drm_decon.c if (atomic_read(&ctx->wait_vsync_event)) { atomic_read 314 drivers/gpu/drm/exynos/exynos_drm_fimd.c !atomic_read(&ctx->wait_vsync_event), atomic_read 953 drivers/gpu/drm/exynos/exynos_drm_fimd.c if (atomic_read(&ctx->triggering)) atomic_read 992 drivers/gpu/drm/exynos/exynos_drm_fimd.c if (atomic_read(&ctx->wait_vsync_event)) { atomic_read 1045 drivers/gpu/drm/exynos/exynos_drm_fimd.c if (atomic_read(&ctx->wait_vsync_event)) { atomic_read 391 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (atomic_read(&g2d_userptr->refcount) > 0) atomic_read 79 drivers/gpu/drm/gma500/mmu.c if (atomic_read(&driver->needs_tlbflush) || force) { atomic_read 110 drivers/gpu/drm/gma500/mmu.c if (atomic_read(&driver->needs_tlbflush)) atomic_read 4282 drivers/gpu/drm/i915/display/intel_display.c if (atomic_read(&dev_priv->gpu_error.pending_fb_pin)) { atomic_read 5594 drivers/gpu/drm/i915/display/intel_dp.c #define C (hdcp->cp_irq_count_cached != atomic_read(&hdcp->cp_irq_count)) atomic_read 5998 drivers/gpu/drm/i915/display/intel_dp.c hdcp->cp_irq_count_cached = atomic_read(&hdcp->cp_irq_count); atomic_read 300 drivers/gpu/drm/i915/display/intel_frontbuffer.c WARN_ON(!(atomic_read(&old->bits) & frontbuffer_bits)); atomic_read 305 drivers/gpu/drm/i915/display/intel_frontbuffer.c WARN_ON(atomic_read(&new->bits) & frontbuffer_bits); atomic_read 83 drivers/gpu/drm/i915/display/intel_frontbuffer.h frontbuffer_bits = atomic_read(&front->bits); atomic_read 111 drivers/gpu/drm/i915/display/intel_frontbuffer.h frontbuffer_bits = atomic_read(&front->bits); atomic_read 202 drivers/gpu/drm/i915/gem/i915_gem_context.c if (atomic_read(&ctx->hw_id_pin_count)) { atomic_read 648 drivers/gpu/drm/i915/gem/i915_gem_context.c GEM_BUG_ON(!atomic_read(&ctx->hw_id_pin_count)); atomic_read 1040 drivers/gpu/drm/i915/gem/i915_gem_context.c return !atomic_read(&ce->pin_count); atomic_read 2073 drivers/gpu/drm/i915/gem/i915_gem_context.c return atomic_read(&file_priv->ban_score) >= I915_CLIENT_SCORE_BANNED; atomic_read 2330 drivers/gpu/drm/i915/gem/i915_gem_context.c args->batch_active = atomic_read(&ctx->guilty_count); atomic_read 2331 drivers/gpu/drm/i915/gem/i915_gem_context.c args->batch_pending = atomic_read(&ctx->active_count); atomic_read 2349 drivers/gpu/drm/i915/gem/i915_gem_context.c GEM_BUG_ON(atomic_read(&ctx->hw_id_pin_count)); atomic_read 2358 drivers/gpu/drm/i915/gem/i915_gem_context.c GEM_BUG_ON(atomic_read(&ctx->hw_id_pin_count) == ~0u); atomic_read 126 drivers/gpu/drm/i915/gem/i915_gem_context.h GEM_BUG_ON(atomic_read(&ctx->hw_id_pin_count) == 0u); atomic_read 223 drivers/gpu/drm/i915/gem/i915_gem_domain.c if (atomic_read(&obj->bind_count)) { atomic_read 124 drivers/gpu/drm/i915/gem/i915_gem_object.c GEM_BUG_ON(!atomic_read(&vma->open_count)); atomic_read 146 drivers/gpu/drm/i915/gem/i915_gem_object.c GEM_BUG_ON(!atomic_read(&i915->mm.free_count)); atomic_read 174 drivers/gpu/drm/i915/gem/i915_gem_object.c GEM_BUG_ON(atomic_read(&obj->bind_count)); atomic_read 270 drivers/gpu/drm/i915/gem/i915_gem_object.h return atomic_read(&obj->mm.pages_pin_count); atomic_read 191 drivers/gpu/drm/i915/gem/i915_gem_pages.c GEM_BUG_ON(atomic_read(&obj->bind_count)); atomic_read 195 drivers/gpu/drm/i915/gem/i915_gem_pages.c if (unlikely(atomic_read(&obj->mm.pages_pin_count))) { atomic_read 72 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (atomic_read(&obj->mm.pages_pin_count) > atomic_read(&obj->bind_count)) atomic_read 244 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c atomic_read(&obj->bind_count)) atomic_read 42 drivers/gpu/drm/i915/gem/selftests/i915_gem_phys.c if (!atomic_read(&obj->mm.pages_pin_count)) { atomic_read 53 drivers/gpu/drm/i915/gt/intel_context.c if (likely(!atomic_read(&ce->pin_count))) { atomic_read 54 drivers/gpu/drm/i915/gt/intel_context.h return atomic_read(&ce->pin_count); atomic_read 1269 drivers/gpu/drm/i915/gt/intel_engine_cs.c enableddisabled(!atomic_read(&engine->execlists.tasklet.count))); atomic_read 1374 drivers/gpu/drm/i915/gt/intel_engine_cs.c drm_printf(m, "\tAwake? %d\n", atomic_read(&engine->wakeref.count)); atomic_read 2379 drivers/gpu/drm/i915/gt/intel_lrc.c atomic_read(&execlists->tasklet.count)); atomic_read 2662 drivers/gpu/drm/i915/gt/intel_lrc.c atomic_read(&execlists->tasklet.count)); atomic_read 74 drivers/gpu/drm/i915/gt/intel_reset.c atomic_read(&file_priv->ban_score)); atomic_read 102 drivers/gpu/drm/i915/gt/intel_reset.c ctx->name, atomic_read(&ctx->guilty_count)); atomic_read 283 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(atomic_read(&timeline->pin_count)); atomic_read 344 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(!atomic_read(&tl->pin_count)); atomic_read 379 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(!atomic_read(&tl->pin_count)); atomic_read 529 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(!atomic_read(&tl->pin_count)); atomic_read 42 drivers/gpu/drm/i915/gt/mock_engine.c GEM_BUG_ON(!atomic_read(&tl->pin_count)); atomic_read 748 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c atomic_read(&execlists->tasklet.count)); atomic_read 924 drivers/gpu/drm/i915/gvt/gtt.c int v = atomic_read(&spt->refcount); atomic_read 932 drivers/gpu/drm/i915/gvt/gtt.c int v = atomic_read(&spt->refcount); atomic_read 1954 drivers/gpu/drm/i915/gvt/gtt.c if (GEM_WARN_ON(atomic_read(&mm->pincount))) atomic_read 2024 drivers/gpu/drm/i915/gvt/gtt.c if (atomic_read(&mm->pincount)) atomic_read 68 drivers/gpu/drm/i915/gvt/scheduler.c if (WARN_ON(!atomic_read(&workload->shadow_mm->pincount))) atomic_read 923 drivers/gpu/drm/i915/gvt/scheduler.c !atomic_read(&workload->shadow_ctx_active)); atomic_read 1076 drivers/gpu/drm/i915/gvt/scheduler.c if (atomic_read(&s->running_workload_num)) { atomic_read 1080 drivers/gpu/drm/i915/gvt/scheduler.c !atomic_read(&s->running_workload_num)); atomic_read 234 drivers/gpu/drm/i915/gvt/vgpu.c if (atomic_read(&vgpu->submission.running_workload_num)) { atomic_read 159 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(!atomic_read(&ref->count)); atomic_read 340 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(!atomic_read(&ref->count)); atomic_read 360 drivers/gpu/drm/i915/i915_active.c if (!atomic_read(&ref->count) && ref->active) atomic_read 490 drivers/gpu/drm/i915/i915_active.c GEM_BUG_ON(atomic_read(&ref->count)); atomic_read 393 drivers/gpu/drm/i915/i915_active.h return !atomic_read(&ref->count); atomic_read 248 drivers/gpu/drm/i915/i915_debugfs.c if (!atomic_read(&obj->bind_count)) atomic_read 361 drivers/gpu/drm/i915/i915_debugfs.c atomic_read(&i915->mm.free_count), atomic_read 659 drivers/gpu/drm/i915/i915_debugfs.c i, atomic_read(®->pin_count)); atomic_read 1583 drivers/gpu/drm/i915/i915_debugfs.c atomic_read(&ctx->hw_id_pin_count)); atomic_read 1734 drivers/gpu/drm/i915/i915_debugfs.c atomic_read(&rps->num_waiters)); atomic_read 1749 drivers/gpu/drm/i915/i915_debugfs.c seq_printf(m, "Wait boosts: %d\n", atomic_read(&rps->boosts)); atomic_read 2327 drivers/gpu/drm/i915/i915_debugfs.c atomic_read(&dev_priv->drm.dev->power.usage_count)); atomic_read 2791 drivers/gpu/drm/i915/i915_debugfs.c atomic_read(&dev_priv->gt.wakeref.count)); atomic_read 2690 drivers/gpu/drm/i915/i915_drv.c WARN_ON_ONCE(atomic_read(&rpm->wakeref_count)); atomic_read 2259 drivers/gpu/drm/i915/i915_drv.h while (atomic_read(&i915->mm.free_count)) { atomic_read 2325 drivers/gpu/drm/i915/i915_drv.h return atomic_read(&error->reset_count); atomic_read 2331 drivers/gpu/drm/i915/i915_drv.h return atomic_read(&error->reset_engine_count[engine->uabi_class]); atomic_read 1683 drivers/gpu/drm/i915/i915_gem.c GEM_BUG_ON(atomic_read(&dev_priv->mm.free_count)); atomic_read 94 drivers/gpu/drm/i915/i915_gem.h return !atomic_read(&t->count); atomic_read 308 drivers/gpu/drm/i915/i915_gem_fence_reg.c if (atomic_read(&fence->pin_count)) atomic_read 321 drivers/gpu/drm/i915/i915_gem_fence_reg.c if (atomic_read(&fence->pin_count)) atomic_read 355 drivers/gpu/drm/i915/i915_gem_fence_reg.c GEM_BUG_ON(atomic_read(&fence->pin_count)); atomic_read 435 drivers/gpu/drm/i915/i915_gem_fence_reg.c count += !atomic_read(&fence->pin_count); atomic_read 776 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(atomic_read(px_used(pd)) > 2 * ARRAY_SIZE(pd->entry)); atomic_read 791 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(atomic_read(px_used(pd)) == 0); atomic_read 975 drivers/gpu/drm/i915/i915_gem_gtt.c idx, len, atomic_read(px_used(pd))); atomic_read 976 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!len || len >= atomic_read(px_used(pd))); atomic_read 1002 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_read(&pt->used)); atomic_read 1003 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!count || count >= atomic_read(&pt->used)); atomic_read 1051 drivers/gpu/drm/i915/i915_gem_gtt.c idx, len, atomic_read(px_used(pd))); atomic_read 1055 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!atomic_read(px_used(pd))); /* Must be pinned! */ atomic_read 1111 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!atomic_read(&pt->used)); atomic_read 1118 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_read(&pt->used)); atomic_read 1122 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(atomic_read(&pt->used) > 2 * I915_PDES); atomic_read 1624 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(count > atomic_read(&pt->used)); atomic_read 1852 drivers/gpu/drm/i915/i915_gem_gtt.c if (px_base(pt) == scratch || atomic_read(&pt->used)) atomic_read 43 drivers/gpu/drm/i915/i915_globals.c if (park.epoch == atomic_read(&epoch)) atomic_read 1268 drivers/gpu/drm/i915/i915_gpu_error.c e->guilty = atomic_read(&ctx->guilty_count); atomic_read 1269 drivers/gpu/drm/i915/i915_gpu_error.c e->active = atomic_read(&ctx->active_count); atomic_read 1626 drivers/gpu/drm/i915/i915_gpu_error.c error->wakelock = atomic_read(&i915->runtime_pm.wakeref_count); atomic_read 1182 drivers/gpu/drm/i915/i915_irq.c client_boost = atomic_read(&rps->num_waiters); atomic_read 292 drivers/gpu/drm/i915/i915_request.c GEM_BUG_ON(!atomic_read(&rq->i915->gt_pm.rps.num_waiters)); atomic_read 97 drivers/gpu/drm/i915/i915_sw_fence.h return atomic_read(&fence->pending) <= 0; atomic_read 102 drivers/gpu/drm/i915/i915_sw_fence.h return atomic_read(&fence->pending) < 0; atomic_read 325 drivers/gpu/drm/i915/i915_sysfs.c boost = atomic_read(&rps->num_waiters); atomic_read 524 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(atomic_read(&obj->mm.pages_pin_count) < atomic_read(&obj->bind_count)); atomic_read 429 drivers/gpu/drm/i915/i915_vma.h GEM_BUG_ON(atomic_read(&vma->fence->pin_count) <= 0); atomic_read 140 drivers/gpu/drm/i915/intel_runtime_pm.c rpm->debug.count, atomic_read(&rpm->wakeref_count))) { atomic_read 598 drivers/gpu/drm/i915/intel_runtime_pm.c int count = atomic_read(&rpm->wakeref_count); atomic_read 121 drivers/gpu/drm/i915/intel_runtime_pm.h __assert_rpm_raw_wakeref_held(rpm, atomic_read(&rpm->wakeref_count)); atomic_read 127 drivers/gpu/drm/i915/intel_runtime_pm.h __assert_rpm_wakelock_held(rpm, atomic_read(&rpm->wakeref_count)); atomic_read 34 drivers/gpu/drm/i915/intel_wakeref.c if (!atomic_read(&wf->count)) { atomic_read 51 drivers/gpu/drm/i915/intel_wakeref.c INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0); atomic_read 121 drivers/gpu/drm/i915/intel_wakeref.h INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0); atomic_read 173 drivers/gpu/drm/i915/intel_wakeref.h INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count)); atomic_read 128 drivers/gpu/drm/i915/selftests/i915_active.c if (atomic_read(&active->base.count) != count) { atomic_read 130 drivers/gpu/drm/i915/selftests/i915_active.c atomic_read(&active->base.count), count); atomic_read 80 drivers/gpu/drm/i915/selftests/i915_gem_evict.c if (atomic_read(&obj->bind_count)) atomic_read 384 drivers/gpu/drm/i915/selftests/i915_request.c atomic_read(&wait->pending), count, atomic_read 44 drivers/gpu/drm/mga/mga_irq.c return atomic_read(&dev_priv->vbl_received); atomic_read 129 drivers/gpu/drm/mga/mga_irq.c (((cur_fence = atomic_read(&dev_priv->last_fence_retired)) atomic_read 155 drivers/gpu/drm/msm/adreno/a5xx_gpu.h int preempt_state = atomic_read(&a5xx_gpu->preempt_state); atomic_read 31 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx])); atomic_read 88 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c enable_count = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx]); atomic_read 122 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c counts = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idxs[0]]); atomic_read 151 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c enable_count = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx]); atomic_read 177 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c counts = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idxs[0]]); atomic_read 268 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx])) atomic_read 320 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c irq_count = atomic_read(&irq_obj->irq_counts[i]); atomic_read 321 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c enable_count = atomic_read(&irq_obj->enable_counts[i]); atomic_read 396 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c if (atomic_read(&dpu_kms->irq_obj.enable_counts[i]) || atomic_read 315 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c if (atomic_read(&dpu_crtc->frame_pending) < 1) { atomic_read 588 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c if (!atomic_read(&dpu_crtc->frame_pending)) { atomic_read 726 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c atomic_read(&dpu_crtc->frame_pending)); atomic_read 731 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c if (atomic_read(&dpu_crtc->frame_pending)) { atomic_read 733 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c atomic_read(&dpu_crtc->frame_pending)); atomic_read 222 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.h return crtc ? atomic_read(&to_dpu_crtc(crtc)->frame_pending) : -EINVAL; atomic_read 262 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(wait_info->atomic_cnt)); atomic_read 280 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(wait_info->atomic_cnt)); atomic_read 292 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(wait_info->atomic_cnt)); atomic_read 299 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(wait_info->atomic_cnt)); atomic_read 1291 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(&phy_enc->underrun_cnt)); atomic_read 1506 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(info->atomic_cnt) == 0, jiffies); atomic_read 1511 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(info->atomic_cnt)); atomic_read 1513 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c } while (atomic_read(info->atomic_cnt) && (rc == 0) && atomic_read 1888 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(&phys->vsync_cnt), atomic_read 1889 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c atomic_read(&phys->underrun_cnt)); atomic_read 217 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c atomic_read(&phys_enc->pending_kickoff_cnt), atomic_read 227 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c atomic_read(&phys_enc->pending_kickoff_cnt)); atomic_read 283 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c refcount = atomic_read(&phys_enc->vblank_refcount); atomic_read 328 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c enable, atomic_read(&phys_enc->vblank_refcount)); atomic_read 601 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c atomic_read(&phys_enc->pending_kickoff_cnt)); atomic_read 618 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c atomic_read(&phys_enc->pending_kickoff_cnt)); atomic_read 309 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c old_cnt = atomic_read(&phys_enc->pending_kickoff_cnt); atomic_read 403 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c refcount = atomic_read(&phys_enc->vblank_refcount); atomic_read 416 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c atomic_read(&phys_enc->vblank_refcount)); atomic_read 651 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c atomic_read(&phys_enc->vblank_refcount)); atomic_read 54 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c if (atomic_read(&hw_blk->refcount)) atomic_read 81 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c atomic_read(&curr->refcount))) atomic_read 127 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c atomic_read(&hw_blk->refcount)); atomic_read 129 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c if (!atomic_read(&hw_blk->refcount)) { atomic_read 64 drivers/gpu/drm/nouveau/nouveau_chan.c if (likely(chan && chan->fence && !atomic_read(&chan->killed))) { atomic_read 502 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c if (retrain || !atomic_read(&dp->lt.done)) atomic_read 546 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c if (atomic_read(&dp->lt.done)) atomic_read 53 drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c args->v0.ref = atomic_read(&chan->ref); atomic_read 340 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c wait_event(clk->wait, !atomic_read(&clk->waiting)); atomic_read 232 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR])) atomic_read 470 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR])) atomic_read 189 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c if (!atomic_read(&vmm->engref[i])) atomic_read 34 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR])) atomic_read 44 drivers/gpu/drm/panfrost/panfrost_gem_shrinker.c if (atomic_read(&bo->gpu_usecount)) atomic_read 591 drivers/gpu/drm/panfrost/panfrost_job.c if (atomic_read(&js->queue[i].sched.hw_rq_count)) atomic_read 171 drivers/gpu/drm/panfrost/panfrost_mmu.c if (!atomic_read(&lru_mmu->as_count)) atomic_read 201 drivers/gpu/drm/panfrost/panfrost_mmu.c WARN_ON(atomic_read(&mmu->as_count) < 0); atomic_read 294 drivers/gpu/drm/qxl/qxl_cmd.c irq_num = atomic_read(&qdev->irq_received_io_cmd); atomic_read 298 drivers/gpu/drm/qxl/qxl_cmd.c atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); atomic_read 301 drivers/gpu/drm/qxl/qxl_cmd.c atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); atomic_read 305 drivers/gpu/drm/qxl/qxl_cmd.c irq_num = atomic_read(&qdev->irq_received_io_cmd); atomic_read 311 drivers/gpu/drm/qxl/qxl_cmd.c atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); atomic_read 314 drivers/gpu/drm/qxl/qxl_cmd.c atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); atomic_read 44 drivers/gpu/drm/qxl/qxl_debugfs.c seq_printf(m, "%d\n", atomic_read(&qdev->irq_received)); atomic_read 45 drivers/gpu/drm/qxl/qxl_debugfs.c seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_display)); atomic_read 46 drivers/gpu/drm/qxl/qxl_debugfs.c seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_cursor)); atomic_read 47 drivers/gpu/drm/qxl/qxl_debugfs.c seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_io_cmd)); atomic_read 47 drivers/gpu/drm/r128/r128_irq.c return atomic_read(&dev_priv->vbl_received); atomic_read 330 drivers/gpu/drm/r128/r128_state.c if (atomic_read(&dev_priv->idle_count) == 0) atomic_read 7076 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 7080 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) { atomic_read 7123 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) { atomic_read 7167 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { atomic_read 7172 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) { atomic_read 7178 drivers/gpu/drm/radeon/cik.c atomic_read(&rdev->irq.pflip[0])) { atomic_read 7183 drivers/gpu/drm/radeon/cik.c atomic_read(&rdev->irq.pflip[1])) { atomic_read 7188 drivers/gpu/drm/radeon/cik.c atomic_read(&rdev->irq.pflip[2])) { atomic_read 7193 drivers/gpu/drm/radeon/cik.c atomic_read(&rdev->irq.pflip[3])) { atomic_read 7198 drivers/gpu/drm/radeon/cik.c atomic_read(&rdev->irq.pflip[4])) { atomic_read 7203 drivers/gpu/drm/radeon/cik.c atomic_read(&rdev->irq.pflip[5])) { atomic_read 7607 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.pflip[0])) atomic_read 7637 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.pflip[1])) atomic_read 7667 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.pflip[2])) atomic_read 7697 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.pflip[3])) atomic_read 7727 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.pflip[4])) atomic_read 7757 drivers/gpu/drm/radeon/cik.c if (atomic_read(&rdev->irq.pflip[5])) atomic_read 4523 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 4527 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) { atomic_read 4531 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) { atomic_read 4536 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 4543 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { atomic_read 4550 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) { atomic_read 4580 drivers/gpu/drm/radeon/evergreen.c atomic_read(&rdev->irq.pflip[i]), "vblank", i); atomic_read 4761 drivers/gpu/drm/radeon/evergreen.c if (atomic_read(&rdev->irq.pflip[crtc_idx])) { atomic_read 721 drivers/gpu/drm/radeon/r100.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 725 drivers/gpu/drm/radeon/r100.c atomic_read(&rdev->irq.pflip[0])) { atomic_read 729 drivers/gpu/drm/radeon/r100.c atomic_read(&rdev->irq.pflip[1])) { atomic_read 794 drivers/gpu/drm/radeon/r100.c if (atomic_read(&rdev->irq.pflip[0])) atomic_read 803 drivers/gpu/drm/radeon/r100.c if (atomic_read(&rdev->irq.pflip[1])) atomic_read 3822 drivers/gpu/drm/radeon/r600.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 3828 drivers/gpu/drm/radeon/r600.c if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { atomic_read 3834 drivers/gpu/drm/radeon/r600.c atomic_read(&rdev->irq.pflip[0])) { atomic_read 3839 drivers/gpu/drm/radeon/r600.c atomic_read(&rdev->irq.pflip[1])) { atomic_read 4146 drivers/gpu/drm/radeon/r600.c if (atomic_read(&rdev->irq.pflip[0])) atomic_read 4176 drivers/gpu/drm/radeon/r600.c if (atomic_read(&rdev->irq.pflip[1])) atomic_read 594 drivers/gpu/drm/radeon/radeon_kms.c *value = atomic_read(&rdev->gpu_reset_counter); atomic_read 257 drivers/gpu/drm/radeon/radeon_ring.c if (rptr != atomic_read(&ring->last_rptr)) { atomic_read 257 drivers/gpu/drm/radeon/radeon_uvd.c uint32_t handle = atomic_read(&rdev->uvd.handles[i]); atomic_read 332 drivers/gpu/drm/radeon/radeon_uvd.c uint32_t handle = atomic_read(&rdev->uvd.handles[i]); atomic_read 517 drivers/gpu/drm/radeon/radeon_uvd.c if (atomic_read(&p->rdev->uvd.handles[i]) == handle) { atomic_read 543 drivers/gpu/drm/radeon/radeon_uvd.c if (atomic_read(&p->rdev->uvd.handles[i]) == handle) { atomic_read 860 drivers/gpu/drm/radeon/radeon_uvd.c if (!atomic_read(&rdev->uvd.handles[i])) atomic_read 205 drivers/gpu/drm/radeon/radeon_vce.c if (atomic_read(&rdev->vce.handles[i])) atomic_read 320 drivers/gpu/drm/radeon/radeon_vce.c uint32_t handle = atomic_read(&rdev->vce.handles[i]); atomic_read 528 drivers/gpu/drm/radeon/radeon_vce.c if (atomic_read(&p->rdev->vce.handles[i]) == handle) { atomic_read 680 drivers/gpu/drm/radeon/rs600.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 684 drivers/gpu/drm/radeon/rs600.c atomic_read(&rdev->irq.pflip[0])) { atomic_read 688 drivers/gpu/drm/radeon/rs600.c atomic_read(&rdev->irq.pflip[1])) { atomic_read 798 drivers/gpu/drm/radeon/rs600.c if (atomic_read(&rdev->irq.pflip[0])) atomic_read 807 drivers/gpu/drm/radeon/rs600.c if (atomic_read(&rdev->irq.pflip[1])) atomic_read 6081 drivers/gpu/drm/radeon/si.c if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) { atomic_read 6085 drivers/gpu/drm/radeon/si.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) { atomic_read 6089 drivers/gpu/drm/radeon/si.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) { atomic_read 6093 drivers/gpu/drm/radeon/si.c if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { atomic_read 6098 drivers/gpu/drm/radeon/si.c if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) { atomic_read 6121 drivers/gpu/drm/radeon/si.c atomic_read(&rdev->irq.pflip[i]), "vblank", i); atomic_read 6304 drivers/gpu/drm/radeon/si.c if (atomic_read(&rdev->irq.pflip[crtc_idx])) { atomic_read 53 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->hw_job_count = atomic_read( atomic_read 147 drivers/gpu/drm/scheduler/sched_entity.c num_jobs = atomic_read(&sched->num_jobs); atomic_read 466 drivers/gpu/drm/scheduler/sched_entity.c if (entity->guilty && atomic_read(entity->guilty)) atomic_read 338 drivers/gpu/drm/scheduler/sched_main.c if (atomic_read(&bad->karma) > atomic_read 487 drivers/gpu/drm/scheduler/sched_main.c if (!found_guilty && atomic_read(&s_job->karma) > sched->hang_limit) { atomic_read 569 drivers/gpu/drm/scheduler/sched_main.c return atomic_read(&sched->hw_rq_count) < atomic_read 123 drivers/gpu/drm/ttm/ttm_bo.c atomic_read(&glob->bo_count)); atomic_read 156 drivers/gpu/drm/ttm/ttm_bo.c BUG_ON(atomic_read(&bo->cpu_writers)); atomic_read 116 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (!ret && unlikely(atomic_read(&bo->cpu_writers) > 0)) { atomic_read 95 drivers/gpu/drm/ttm/ttm_module.c wait_event(exit_q, atomic_read(&device_released) == 1); atomic_read 90 drivers/gpu/drm/vc4/vc4_hvs.c drm_printf(&p, "%d\n", atomic_read(&vc4->underrun)); atomic_read 100 drivers/gpu/drm/via/via_irq.c return atomic_read(&dev_priv->vbl_received); atomic_read 116 drivers/gpu/drm/via/via_irq.c if (!(atomic_read(&dev_priv->vbl_received) & 0x0F)) { atomic_read 126 drivers/gpu/drm/via/via_irq.c if (!(atomic_read(&dev_priv->vbl_received) & 0xFF)) { atomic_read 241 drivers/gpu/drm/via/via_irq.c cur_irq_sequence = atomic_read(&cur_irq->irq_received); atomic_read 245 drivers/gpu/drm/via/via_irq.c atomic_read(&cur_irq->irq_received)) - atomic_read 365 drivers/gpu/drm/via/via_irq.c atomic_read(&cur_irq->irq_received); atomic_read 541 drivers/gpu/drm/virtio/virtgpu_ioctl.c atomic_read(&cache_ent->is_valid), 5 * HZ); atomic_read 1312 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c if (atomic_read(&dev_priv->num_fifo_resources) != 0) { atomic_read 547 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c *seqno = atomic_read(&dev_priv->marker_seq); atomic_read 153 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c ret = ((atomic_read(&dev_priv->marker_seq) - seqno) atomic_read 192 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c signal_seq = atomic_read(&dev_priv->marker_seq); atomic_read 138 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c seqno = atomic_read(&dev_priv->marker_seq); atomic_read 314 drivers/gpu/host1x/syncpt.c current_val = (u32)atomic_read(&sp->min_val); atomic_read 315 drivers/gpu/host1x/syncpt.c future_val = (u32)atomic_read(&sp->max_val); atomic_read 478 drivers/gpu/host1x/syncpt.c return (u32)atomic_read(&sp->max_val); atomic_read 493 drivers/gpu/host1x/syncpt.c return (u32)atomic_read(&sp->min_val); atomic_read 87 drivers/gpu/host1x/syncpt.h min = atomic_read(&sp->min_val); atomic_read 88 drivers/gpu/host1x/syncpt.h max = atomic_read(&sp->max_val); atomic_read 112 drivers/greybus/greybus_trace.h __entry->waiters = atomic_read(&operation->waiters); atomic_read 90 drivers/greybus/operation.c if (atomic_read(&operation->waiters)) atomic_read 370 drivers/hid/hid-cp2112.c atomic_read(avail), msecs_to_jiffies(RESPONSE_TIMEOUT)); atomic_read 1838 drivers/hid/hid-logitech-hidpp.c s = atomic_read(&data->workqueue_size); atomic_read 3354 drivers/hid/hid-logitech-hidpp.c bool connected = atomic_read(&hidpp->connected); atomic_read 292 drivers/hsi/clients/ssi_protocol.c if (atomic_read(&ssi->tx_usecnt)) atomic_read 346 drivers/hsi/clients/ssi_protocol.c dev_dbg(&master->device, "start TX %d\n", atomic_read(&ssi->tx_usecnt)); atomic_read 363 drivers/hsi/clients/ssi_protocol.c WARN_ON_ONCE(atomic_read(&ssi->tx_usecnt) == 0); atomic_read 374 drivers/hsi/clients/ssi_protocol.c dev_dbg(&master->device, "stop TX %d\n", atomic_read(&ssi->tx_usecnt)); atomic_read 467 drivers/hsi/clients/ssi_protocol.c if (atomic_read(&ssi->tx_usecnt) == 0) atomic_read 863 drivers/hsi/clients/ssi_protocol.c if (atomic_read(&ssi->tx_usecnt)) { atomic_read 1038 drivers/hv/channel_mgmt.c while (atomic_read(&vmbus_connection.offer_in_progress) != 0) { atomic_read 1172 drivers/hv/hv_balloon.c if (status.hdr.trans_id != atomic_read(&trans_id)) atomic_read 2182 drivers/hv/vmbus_drv.c while (atomic_read(&vmbus_connection.offer_in_progress) != 0) { atomic_read 2213 drivers/hv/vmbus_drv.c if (atomic_read(&vmbus_connection.nr_chan_close_on_suspend) > 0) atomic_read 2216 drivers/hv/vmbus_drv.c WARN_ON(atomic_read(&vmbus_connection.nr_chan_fixup_on_resume) != 0); atomic_read 2286 drivers/hv/vmbus_drv.c WARN_ON(atomic_read(&vmbus_connection.nr_chan_fixup_on_resume) == 0); atomic_read 60 drivers/hwmon/pwm-fan.c pulses = atomic_read(&ctx->pulses); atomic_read 539 drivers/hwmon/sht15.c if (!atomic_read(&data->interrupt_handled)) atomic_read 831 drivers/hwmon/sht15.c || atomic_read(&data->interrupt_handled)) atomic_read 446 drivers/hwtracing/coresight/coresight-etb10.c if (atomic_read(csdev->refcnt) != 1) atomic_read 85 drivers/hwtracing/coresight/coresight-funnel.c if (atomic_read(&csdev->refcnt[inport]) == 0) { atomic_read 106 drivers/hwtracing/coresight/coresight-replicator.c if (atomic_read(&csdev->refcnt[outport]) == 0) { atomic_read 348 drivers/hwtracing/coresight/coresight-tmc-etf.c if (atomic_read(&csdev->refcnt[0]) == 0) { atomic_read 464 drivers/hwtracing/coresight/coresight-tmc-etf.c if (atomic_read(csdev->refcnt) != 1) atomic_read 1489 drivers/hwtracing/coresight/coresight-tmc-etr.c if (atomic_read(csdev->refcnt) != 1) { atomic_read 307 drivers/hwtracing/coresight/coresight.c if (atomic_read(&csdev->refcnt[i]) != 0) atomic_read 1287 drivers/hwtracing/intel_th/msu.c if (atomic_read(&msc->user_count) != -1) atomic_read 2570 drivers/i3c/master.c if (atomic_read(&dev->ibi->pending_ibis)) atomic_read 239 drivers/iio/accel/hid-sensor-accel-3d.c if (atomic_read(&accel_state->common_attributes.data_ready)) { atomic_read 169 drivers/iio/adc/stm32-dfsdm-core.c atomic_read(&priv->n_active_ch)); atomic_read 211 drivers/iio/adc/stm32-dfsdm-core.c atomic_read(&priv->n_active_ch)); atomic_read 358 drivers/iio/adc/sun4i-gpadc-iio.c if (atomic_read(&info->ignore_temp_data_irq)) atomic_read 372 drivers/iio/adc/sun4i-gpadc-iio.c if (atomic_read(&info->ignore_fifo_data_irq)) atomic_read 215 drivers/iio/common/hid-sensors/hid-sensor-trigger.c if (atomic_read(&attrb->user_requested_state)) atomic_read 227 drivers/iio/common/hid-sensors/hid-sensor-trigger.c if (atomic_read(&attrb->runtime_pm_enable)) atomic_read 633 drivers/iio/common/ssp_sensors/ssp_dev.c if (atomic_read(&data->enable_refcount) > 0) atomic_read 658 drivers/iio/common/ssp_sensors/ssp_dev.c if (atomic_read(&data->enable_refcount) > 0) atomic_read 200 drivers/iio/gyro/hid-sensor-gyro-3d.c if (atomic_read(&gyro_state->common_attributes.data_ready)) atomic_read 129 drivers/iio/humidity/hid-sensor-humidity.c if (atomic_read(&humid_st->common_attributes.data_ready)) atomic_read 166 drivers/iio/industrialio-trigger.c if (!atomic_read(&trig->use_count)) { atomic_read 190 drivers/iio/industrialio-trigger.c if (!atomic_read(&trig->use_count)) { atomic_read 190 drivers/iio/light/hid-sensor-als.c if (atomic_read(&als_state->common_attributes.data_ready)) atomic_read 170 drivers/iio/light/hid-sensor-prox.c if (atomic_read(&prox_state->common_attributes.data_ready)) atomic_read 110 drivers/iio/light/lm3533-als.c *zone = atomic_read(&als->zone); atomic_read 294 drivers/iio/magnetometer/hid-sensor-magn-3d.c if (atomic_read(&magn_state->magn_flux_attributes.data_ready)) atomic_read 199 drivers/iio/orientation/hid-sensor-incl-3d.c if (atomic_read(&incl_state->common_attributes.data_ready)) atomic_read 155 drivers/iio/orientation/hid-sensor-rotation.c if (atomic_read(&rot_state->common_attributes.data_ready)) atomic_read 174 drivers/iio/pressure/hid-sensor-press.c if (atomic_read(&press_state->common_attributes.data_ready)) atomic_read 127 drivers/iio/temperature/hid-sensor-temperature.c if (atomic_read(&temp_st->common_attributes.data_ready)) atomic_read 290 drivers/infiniband/core/cq.c if (WARN_ON_ONCE(atomic_read(&cq->usecnt))) atomic_read 182 drivers/infiniband/core/fmr_pool.c if (atomic_read(&pool->flush_ser) - atomic_read(&pool->req_ser) < 0) atomic_read 378 drivers/infiniband/core/fmr_pool.c atomic_read(&pool->flush_ser) - serial >= 0)) atomic_read 214 drivers/infiniband/core/iwcm.c BUG_ON(atomic_read(&cm_id_priv->refcount)==0); atomic_read 90 drivers/infiniband/core/iwpm_msg.c msg_seq = atomic_read(&echo_nlmsg_seq); atomic_read 177 drivers/infiniband/core/iwpm_msg.c msg_seq = atomic_read(&echo_nlmsg_seq); atomic_read 266 drivers/infiniband/core/iwpm_msg.c msg_seq = atomic_read(&echo_nlmsg_seq); atomic_read 350 drivers/infiniband/core/iwpm_msg.c msg_seq = atomic_read(&echo_nlmsg_seq); atomic_read 64 drivers/infiniband/core/iwpm_util.c if (atomic_read(&iwpm_admin.refcount) == 0) { atomic_read 108 drivers/infiniband/core/iwpm_util.c if (atomic_read(&iwpm_admin.refcount) == 0) { atomic_read 2311 drivers/infiniband/core/mad.c if (atomic_read(&qp_info->snoop_count)) atomic_read 2560 drivers/infiniband/core/mad.c if (atomic_read(&qp_info->snoop_count)) atomic_read 2804 drivers/infiniband/core/mad.c if (atomic_read(&recv_mad_agent->qp_info->snoop_count)) atomic_read 2822 drivers/infiniband/core/mad.c if (atomic_read(&mad_agent_priv->qp_info->snoop_count)) atomic_read 549 drivers/infiniband/core/nldev.c atomic_read(&cq->usecnt), RDMA_NLDEV_ATTR_PAD)) atomic_read 629 drivers/infiniband/core/nldev.c atomic_read(&pd->usecnt), RDMA_NLDEV_ATTR_PAD)) atomic_read 99 drivers/infiniband/core/rdma_core.c WARN_ON(atomic_read(&uobj->usecnt) <= 0); atomic_read 102 drivers/infiniband/core/rdma_core.c WARN_ON(atomic_read(&uobj->usecnt) != -1); atomic_read 210 drivers/infiniband/core/security.c if (atomic_read(&pp->sec->error_list_count)) atomic_read 476 drivers/infiniband/core/security.c sec->error_comps_pending = atomic_read(&sec->error_list_count); atomic_read 98 drivers/infiniband/core/uverbs_std_types_counters.c if (!atomic_read(&counters->usecnt)) atomic_read 340 drivers/infiniband/core/verbs.c WARN_ON(atomic_read(&pd->usecnt)); atomic_read 1031 drivers/infiniband/core/verbs.c if (atomic_read(&srq->usecnt)) atomic_read 1839 drivers/infiniband/core/verbs.c if (atomic_read(&real_qp->usecnt) == 0) atomic_read 1869 drivers/infiniband/core/verbs.c if (atomic_read(&qp->usecnt)) atomic_read 1962 drivers/infiniband/core/verbs.c if (atomic_read(&cq->usecnt)) atomic_read 2244 drivers/infiniband/core/verbs.c if (atomic_read(&xrcd->usecnt)) atomic_read 2308 drivers/infiniband/core/verbs.c if (atomic_read(&wq->usecnt)) atomic_read 2389 drivers/infiniband/core/verbs.c if (atomic_read(&rwq_ind_table->usecnt)) atomic_read 126 drivers/infiniband/hw/bnxt_re/hw_counters.c stats->value[BNXT_RE_ACTIVE_QP] = atomic_read(&rdev->qp_count); atomic_read 127 drivers/infiniband/hw/bnxt_re/hw_counters.c stats->value[BNXT_RE_ACTIVE_SRQ] = atomic_read(&rdev->srq_count); atomic_read 128 drivers/infiniband/hw/bnxt_re/hw_counters.c stats->value[BNXT_RE_ACTIVE_CQ] = atomic_read(&rdev->cq_count); atomic_read 129 drivers/infiniband/hw/bnxt_re/hw_counters.c stats->value[BNXT_RE_ACTIVE_MR] = atomic_read(&rdev->mr_count); atomic_read 130 drivers/infiniband/hw/bnxt_re/hw_counters.c stats->value[BNXT_RE_ACTIVE_MW] = atomic_read(&rdev->mw_count); atomic_read 1648 drivers/infiniband/hw/bnxt_re/main.c if (atomic_read(&rdev->sched_count) > 0) atomic_read 163 drivers/infiniband/hw/bnxt_re/qplib_fp.c if (atomic_read(&cq->arm_state) && nq->cqn_handler) { atomic_read 100 drivers/infiniband/hw/cxgb3/iwch_provider.c wait_event(chp->wait, !atomic_read(&chp->refcnt)); atomic_read 682 drivers/infiniband/hw/cxgb3/iwch_provider.c wait_event(qhp->wait, !atomic_read(&qhp->refcnt)); atomic_read 980 drivers/infiniband/hw/cxgb4/cq.c wait_event(chp->wait, !atomic_read(&chp->refcnt)); atomic_read 132 drivers/infiniband/hw/cxgb4/device.c idx = atomic_read(&dev->rdev.wr_log_idx) & atomic_read 1279 drivers/infiniband/hw/hfi1/driver.c if (atomic_read(&ppd->led_override_timer_active)) { atomic_read 325 drivers/infiniband/hw/hfi1/file_ops.c if (atomic_read(&pq->n_reqs) == pq->n_max_reqs) { atomic_read 217 drivers/infiniband/hw/hfi1/iowait.h wait_event(wait->wait_dma, !atomic_read(&wait->sdma_busy)); atomic_read 228 drivers/infiniband/hw/hfi1/iowait.h return atomic_read(&wait->sdma_busy); atomic_read 271 drivers/infiniband/hw/hfi1/iowait.h !atomic_read(&wait->pio_busy), atomic_read 283 drivers/infiniband/hw/hfi1/iowait.h return atomic_read(&wait->pio_busy); atomic_read 878 drivers/infiniband/hw/hfi1/mad.c is_beaconing_active = !!atomic_read(&ppd->led_override_timer_active); atomic_read 4060 drivers/infiniband/hw/hfi1/mad.c is_beaconing_active = !!atomic_read(&ppd->led_override_timer_active); atomic_read 671 drivers/infiniband/hw/hfi1/qp.c atomic_read(&qp->refcount), atomic_read 668 drivers/infiniband/hw/hfi1/rc.c if (atomic_read(&priv->n_tid_requests) >= atomic_read 3369 drivers/infiniband/hw/hfi1/sdma.c atomic_read(&dd->sdma_unfreeze_count) <= atomic_read 3372 drivers/infiniband/hw/hfi1/sdma.c if (ret || atomic_read(&dd->sdma_unfreeze_count) < 0) atomic_read 3388 drivers/infiniband/hw/hfi1/sdma.c atomic_read(&dd->sdma_unfreeze_count) <= 0); atomic_read 5014 drivers/infiniband/hw/hfi1/tid_rdma.c if (((atomic_read(&priv->n_tid_requests) < HFI1_TID_RDMA_WRITE_CNT) && atomic_read 5015 drivers/infiniband/hw/hfi1/tid_rdma.c atomic_read(&priv->n_requests) && atomic_read 1500 drivers/infiniband/hw/hfi1/trace_tid.h __entry->n_requests = atomic_read(&priv->n_requests); atomic_read 1501 drivers/infiniband/hw/hfi1/trace_tid.h __entry->n_tid_requests = atomic_read(&priv->n_tid_requests); atomic_read 297 drivers/infiniband/hw/hfi1/user_sdma.c !atomic_read(&pq->n_reqs)); atomic_read 984 drivers/infiniband/hw/hfi1/user_sdma.c atomic_read(&pq->n_locked), npages)) { atomic_read 1514 drivers/infiniband/hw/hfi1/user_sdma.c if (atomic_read(&node->refcount)) atomic_read 1541 drivers/infiniband/hw/hfi1/user_sdma.c if (!atomic_read(&node->refcount)) atomic_read 1948 drivers/infiniband/hw/i40iw/i40iw_cm.c if (atomic_read(&listener->pend_accepts_cnt) > 0) atomic_read 1954 drivers/infiniband/hw/i40iw/i40iw_cm.c atomic_read(&listener->pend_accepts_cnt)); atomic_read 2598 drivers/infiniband/hw/i40iw/i40iw_cm.c if (atomic_read(&cm_node->listener->pend_accepts_cnt) > atomic_read 362 drivers/infiniband/hw/i40iw/i40iw_hw.c if (atomic_read(&iwqp->close_timer_started)) atomic_read 1763 drivers/infiniband/hw/i40iw/i40iw_main.c if (atomic_read(&iwdev->params_busy)) atomic_read 1977 drivers/infiniband/hw/i40iw/i40iw_main.c (atomic_read(&iwdev->vchnl_msgs) == 0)) atomic_read 1984 drivers/infiniband/hw/i40iw/i40iw_main.c (atomic_read(&iwdev->vchnl_msgs) == 0), atomic_read 449 drivers/infiniband/hw/i40iw/i40iw_utils.c !atomic_read(&cqp_request->refcount), atomic_read 1170 drivers/infiniband/hw/i40iw/i40iw_utils.c (atomic_read(&iwdev->vchnl_msgs) == 1), atomic_read 1017 drivers/infiniband/hw/mlx4/mcg.c atomic_read(&group->refcount), atomic_read 1104 drivers/infiniband/hw/mlx4/mcg.c if (atomic_read(&group->refcount)) atomic_read 1106 drivers/infiniband/hw/mlx4/mcg.c atomic_read(&group->refcount), group); atomic_read 1234 drivers/infiniband/hw/mlx4/mcg.c if (atomic_read(&group->refcount)) { atomic_read 3210 drivers/infiniband/hw/mlx5/main.c atomic_read(&handler->ibcounters->usecnt) == 1) atomic_read 3641 drivers/infiniband/hw/mlx5/main.c atomic_read(&handler->ibcounters->usecnt) == 1) atomic_read 4408 drivers/infiniband/hw/mlx5/main.c return sprintf(buf, "%d\n", atomic_read(&dev->mdev->priv.reg_pages)); atomic_read 1590 drivers/infiniband/hw/mlx5/mr.c if (atomic_read(&mr->num_pending_prefetch)) atomic_read 1592 drivers/infiniband/hw/mlx5/mr.c WARN_ON(atomic_read(&mr->num_pending_prefetch)); atomic_read 606 drivers/infiniband/hw/mlx5/odp.c wait_event(imr->q_leaf_free, !atomic_read(&imr->num_leaf_free)); atomic_read 170 drivers/infiniband/hw/mthca/mthca_av.c } else if (!atomic_read(&pd->sqp_count) && atomic_read 774 drivers/infiniband/hw/qib/qib_driver.c if (atomic_read(&ppd->led_override_timer_active)) { atomic_read 359 drivers/infiniband/hw/qib/qib_qp.c wait_event(priv->wait_dma, !atomic_read(&priv->s_dma_busy)); atomic_read 440 drivers/infiniband/hw/qib/qib_qp.c atomic_read(&priv->s_dma_busy), atomic_read 247 drivers/infiniband/hw/qib/qib_rc.c if (atomic_read(&priv->s_dma_busy)) { atomic_read 66 drivers/infiniband/hw/qib/qib_uc.c if (atomic_read(&priv->s_dma_busy)) { atomic_read 253 drivers/infiniband/hw/qib/qib_ud.c if (atomic_read(&priv->s_dma_busy)) { atomic_read 293 drivers/infiniband/hw/qib/qib_ud.c if (atomic_read(&priv->s_dma_busy)) { atomic_read 71 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h const unsigned int idx = atomic_read(var); atomic_read 80 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h __u32 idx = atomic_read(var) + 1; /* Increment. */ atomic_read 89 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h const __u32 tail = atomic_read(&r->prod_tail); atomic_read 90 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h const __u32 head = atomic_read(&r->cons_head); atomic_read 103 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h const __u32 tail = atomic_read(&r->prod_tail); atomic_read 104 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h const __u32 head = atomic_read(&r->cons_head); atomic_read 413 drivers/infiniband/sw/rdmavt/mcast.c wait_event(mcast->wait, atomic_read(&mcast->refcount) <= 1); atomic_read 418 drivers/infiniband/sw/rdmavt/mcast.c wait_event(mcast->wait, !atomic_read(&mcast->refcount)); atomic_read 958 drivers/infiniband/sw/rdmavt/mr.c if (unlikely(atomic_read(&mr->lkey_invalid) || atomic_read 1066 drivers/infiniband/sw/rdmavt/mr.c if (unlikely(atomic_read(&mr->lkey_invalid) || atomic_read 304 drivers/infiniband/sw/rdmavt/qp.c return atomic_read(&wss->total_count) >= wss->threshold; atomic_read 1724 drivers/infiniband/sw/rdmavt/qp.c wait_event(qp->wait, !atomic_read(&qp->refcount)); atomic_read 1962 drivers/infiniband/sw/rdmavt/qp.c reserved_used = atomic_read(&qp->s_reserved_used); atomic_read 1977 drivers/infiniband/sw/rdmavt/qp.c reserved_used = atomic_read(&qp->s_reserved_used); atomic_read 2051 drivers/infiniband/sw/rdmavt/qp.c atomic_read(&qp->local_ops_pending)) { atomic_read 260 drivers/infiniband/sw/rxe/rxe_pool.c if (atomic_read(&pool->num_elem) > 0) atomic_read 673 drivers/infiniband/sw/rxe/rxe_req.c if (unlikely(atomic_read(&qp->skb_out) > atomic_read 971 drivers/infiniband/sw/siw/siw_cm.c if (atomic_read(&new_s->sk->sk_rmem_alloc)) { atomic_read 214 drivers/infiniband/sw/siw/siw_main.c usage = atomic_read(&per_cpu(siw_use_cnt, cpu)); atomic_read 123 drivers/infiniband/sw/siw/siw_verbs.c atomic_read(&sdev->num_ctx)); atomic_read 130 drivers/infiniband/sw/siw/siw_verbs.c atomic_read(&sdev->num_ctx)); atomic_read 273 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "now %d PD's(s)\n", atomic_read(&sdev->num_pd)); atomic_read 1613 drivers/infiniband/ulp/ipoib/ipoib_main.c wait_flushed = atomic_read(&priv->ntbl.entries); atomic_read 2763 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_read(&ch->sq_wr_avail)); atomic_read 2610 drivers/input/mouse/cyapa_gen5.c if (atomic_read(&pip->cmd_issued)) { atomic_read 1078 drivers/input/touchscreen/cyttsp4_core.c if (atomic_read(&cd->ignore_irq)) { atomic_read 269 drivers/iommu/amd_iommu_v2.c wait_event(pasid_state->wq, !atomic_read(&pasid_state->count)); atomic_read 856 drivers/iommu/amd_iommu_v2.c wait_event(dev_state->wq, !atomic_read(&dev_state->count)); atomic_read 1031 drivers/iommu/arm-smmu-v3.c if (atomic_read(&cmdq->lock) == 1) atomic_read 1946 drivers/iommu/arm-smmu-v3.c if (!atomic_read(&smmu_domain->nr_ats_masters)) atomic_read 464 drivers/iommu/intel-svm.c else if (atomic_read(&svm->mm->mm_users) > 0) atomic_read 581 drivers/iommu/iova.c if (!atomic_read(&iovad->fq_timer_on) && atomic_read 579 drivers/irqchip/irq-renesas-intc-irqpin.c if (atomic_read(&p->wakeup_path)) atomic_read 233 drivers/irqchip/irq-renesas-irqc.c if (atomic_read(&p->wakeup_path)) atomic_read 79 drivers/leds/trigger/ledtrig-cpu.c active_cpus = atomic_read(&num_active_cpus); atomic_read 258 drivers/leds/trigger/ledtrig-netdev.c jiffies_to_msecs(atomic_read(&trigger_data->interval))); atomic_read 379 drivers/leds/trigger/ledtrig-netdev.c atomic_read(&trigger_data->interval)); atomic_read 389 drivers/leds/trigger/ledtrig-netdev.c (atomic_read(&trigger_data->interval)*2)); atomic_read 388 drivers/lightnvm/pblk-core.c if (!atomic_read(&pblk->inflight_io)) atomic_read 1138 drivers/lightnvm/pblk-core.c int blk_to_erase = atomic_read(&line->blk_in_line); atomic_read 1160 drivers/lightnvm/pblk-core.c int blk_in_line = atomic_read(&line->blk_in_line); atomic_read 1561 drivers/lightnvm/pblk-core.c left_seblks = atomic_read(&new->left_seblks); atomic_read 1564 drivers/lightnvm/pblk-core.c if (atomic_read(&new->left_eblks)) { atomic_read 380 drivers/lightnvm/pblk-gc.c if (!atomic_read(&line->sec_to_update)) atomic_read 397 drivers/lightnvm/pblk-gc.c unsigned int werr_lines = atomic_read(&rl->werr_lines); atomic_read 456 drivers/lightnvm/pblk-gc.c if (!run_gc || (atomic_read(&gc->read_inflight_gc) >= PBLK_GC_L_QD)) atomic_read 549 drivers/lightnvm/pblk-gc.c atomic_read(&gc->pipeline_gc)); atomic_read 553 drivers/lightnvm/pblk-gc.c if (!atomic_read(&gc->pipeline_gc)) atomic_read 830 drivers/lightnvm/pblk-rb.c atomic_read(&rb->inflight_flush_point), atomic_read 848 drivers/lightnvm/pblk-rb.c atomic_read(&rb->inflight_flush_point), atomic_read 31 drivers/lightnvm/pblk-rl.c rb_space = atomic_read(&rl->rb_space); atomic_read 38 drivers/lightnvm/pblk-rl.c int rb_user_cnt = atomic_read(&rl->rb_user_cnt); atomic_read 39 drivers/lightnvm/pblk-rl.c int rb_space = atomic_read(&rl->rb_space); atomic_read 52 drivers/lightnvm/pblk-rl.c int rb_space = atomic_read(&rl->rb_space); atomic_read 60 drivers/lightnvm/pblk-rl.c int rb_gc_cnt = atomic_read(&rl->rb_gc_cnt); atomic_read 100 drivers/lightnvm/pblk-rl.c return atomic_read(&rl->free_blocks); atomic_read 105 drivers/lightnvm/pblk-rl.c return atomic_read(&rl->free_user_blocks); atomic_read 113 drivers/lightnvm/pblk-rl.c int werr_gc_needed = atomic_read(&rl->werr_lines); atomic_read 161 drivers/lightnvm/pblk-rl.c int blk_in_line = atomic_read(&line->blk_in_line); atomic_read 173 drivers/lightnvm/pblk-rl.c int blk_in_line = atomic_read(&line->blk_in_line); atomic_read 182 drivers/lightnvm/pblk-rl.c free_blocks = atomic_read(&rl->free_user_blocks); atomic_read 60 drivers/lightnvm/pblk-sysfs.c rb_user_cnt = atomic_read(&pblk->rl.rb_user_cnt); atomic_read 62 drivers/lightnvm/pblk-sysfs.c rb_gc_cnt = atomic_read(&pblk->rl.rb_gc_cnt); atomic_read 298 drivers/lightnvm/pblk-sysfs.c atomic_read(&pblk->gc.read_inflight_gc)); atomic_read 304 drivers/lightnvm/pblk-sysfs.c atomic_read(&pblk->inflight_io)); atomic_read 338 drivers/lightnvm/pblk-write.c if (likely(!e_line || !atomic_read(&e_line->left_eblks))) atomic_read 627 drivers/macintosh/adb.c if (atomic_read(&state->n_pending) == 0) { atomic_read 701 drivers/macintosh/adb.c if (atomic_read(&state->n_pending) == 0 atomic_read 737 drivers/macintosh/adb.c else if (atomic_read(&state->n_pending) == 0) atomic_read 980 drivers/mailbox/bcm-flexrm-mailbox.c (u32)atomic_read(&ring->msg_send_count), atomic_read 981 drivers/mailbox/bcm-flexrm-mailbox.c (u32)atomic_read(&ring->msg_cmpl_count)); atomic_read 97 drivers/md/bcache/alloc.c r = atomic_read(&c->rescale); atomic_read 111 drivers/md/bcache/alloc.c !atomic_read(&b->pin)) { atomic_read 137 drivers/md/bcache/alloc.c !atomic_read(&b->pin) && atomic_read 364 drivers/md/bcache/alloc.c allocator_wait(ca, !atomic_read(&ca->set->prio_blocked)); atomic_read 448 drivers/md/bcache/alloc.c BUG_ON(atomic_read(&b->pin) != 1); atomic_read 1675 drivers/md/bcache/btree.c if (atomic_read(&b->c->search_inflight) && atomic_read 1752 drivers/md/bcache/btree.c if (!atomic_read(&b->pin)) { atomic_read 1815 drivers/md/bcache/btree.c if (atomic_read(&b->pin)) atomic_read 1881 drivers/md/bcache/btree.c if (atomic_read(&c->sectors_to_gc) < 0) atomic_read 90 drivers/md/bcache/closure.c if (atomic_read(&cl->remaining) & CLOSURE_WAITING) atomic_read 176 drivers/md/bcache/closure.c int r = atomic_read(&cl->remaining); atomic_read 183 drivers/md/bcache/closure.h if ((atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK) != 1) atomic_read 200 drivers/md/bcache/extents.c buf, PTR_BUCKET_NR(b->c, k, i), atomic_read(&g->pin), atomic_read 532 drivers/md/bcache/extents.c buf, PTR_BUCKET_NR(b->c, k, ptr), atomic_read(&g->pin), atomic_read 109 drivers/md/bcache/io.c errors = atomic_read(&ca->io_errors); atomic_read 150 drivers/md/bcache/io.c int congested = atomic_read(&c->congested); atomic_read 369 drivers/md/bcache/journal.c BUG_ON(i->pin && atomic_read(i->pin) != 1); atomic_read 444 drivers/md/bcache/journal.c ref_nr = atomic_read(fifo_front_p); atomic_read 481 drivers/md/bcache/journal.c ref_nr = atomic_read(fifo_front_p); atomic_read 608 drivers/md/bcache/journal.c switch (atomic_read(&ja->discard_in_flight)) { atomic_read 649 drivers/md/bcache/journal.c while (!atomic_read(&fifo_front(&c->journal.pin))) atomic_read 219 drivers/md/bcache/movinggc.c atomic_read(&b->pin)) atomic_read 73 drivers/md/bcache/request.c while (atomic_read(&s->cl.remaining) & CLOSURE_WAITING) atomic_read 348 drivers/md/bcache/request.c i += atomic_read(&c->congested); atomic_read 620 drivers/md/bcache/request.c if (dc && atomic_read(&dc->has_dirty)) atomic_read 1192 drivers/md/bcache/request.c if (atomic_read(&d->c->idle_counter)) atomic_read 1200 drivers/md/bcache/request.c if (unlikely(atomic_read(&d->c->at_max_writeback_rate) == 1)) { atomic_read 182 drivers/md/bcache/stats.c if (!atomic_read(&acc->closing)) atomic_read 1270 drivers/md/bcache/super.c if (atomic_read(&dc->running)) atomic_read 1678 drivers/md/bcache/super.c } else if (atomic_read(&dc->has_dirty)) { atomic_read 197 drivers/md/bcache/sysfs.c sysfs_printf(io_errors, "%i", atomic_read(&dc->io_errors)); atomic_read 252 drivers/md/bcache/sysfs.c sysfs_print(running, atomic_read(&dc->running)); atomic_read 1009 drivers/md/bcache/sysfs.c atomic_read(&ca->io_errors) >> IO_ERROR_SHIFT); atomic_read 142 drivers/md/bcache/writeback.c atomic_read(&c->attached_dev_nr) * 6) atomic_read 145 drivers/md/bcache/writeback.c if (atomic_read(&c->at_max_writeback_rate) != 1) atomic_read 161 drivers/md/bcache/writeback.c if ((atomic_read(&c->idle_counter) < atomic_read 162 drivers/md/bcache/writeback.c atomic_read(&c->attached_dev_nr) * 6) || atomic_read 163 drivers/md/bcache/writeback.c !atomic_read(&c->at_max_writeback_rate)) atomic_read 196 drivers/md/bcache/writeback.c if (atomic_read(&dc->has_dirty) && dc->writeback_percent) { atomic_read 330 drivers/md/bcache/writeback.c if (atomic_read(&dc->writeback_sequence_next) != io->sequence) { atomic_read 334 drivers/md/bcache/writeback.c if (atomic_read(&dc->writeback_sequence_next) == io->sequence) { atomic_read 674 drivers/md/bcache/writeback.c (!atomic_read(&dc->has_dirty) || !dc->writeback_running)) { atomic_read 775 drivers/md/bcache/writeback.c if (atomic_read(&b->c->search_inflight) && atomic_read 31 drivers/md/bcache/writeback.h ret += atomic_read(d->stripe_sectors_dirty + i); atomic_read 50 drivers/md/bcache/writeback.h if (atomic_read(dc->disk.stripe_sectors_dirty + stripe)) atomic_read 98 drivers/md/bcache/writeback.h if (!atomic_read(&dc->has_dirty) && atomic_read 38 drivers/md/dm-bio-record.h bd->__bi_remaining = atomic_read(&bio->__bi_remaining); atomic_read 152 drivers/md/dm-cache-background-tracker.c return atomic_read(&b->pending_writebacks); atomic_read 158 drivers/md/dm-cache-background-tracker.c return atomic_read(&b->pending_demotes); atomic_read 164 drivers/md/dm-cache-background-tracker.c return atomic_read(&b->pending_promotes) + atomic_read 165 drivers/md/dm-cache-background-tracker.c atomic_read(&b->pending_writebacks) + atomic_read 166 drivers/md/dm-cache-background-tracker.c atomic_read(&b->pending_demotes) >= b->max_work; atomic_read 1054 drivers/md/dm-cache-target.c stats.read_hits = atomic_read(&cache->stats.read_hit); atomic_read 1055 drivers/md/dm-cache-target.c stats.read_misses = atomic_read(&cache->stats.read_miss); atomic_read 1056 drivers/md/dm-cache-target.c stats.write_hits = atomic_read(&cache->stats.write_hit); atomic_read 1057 drivers/md/dm-cache-target.c stats.write_misses = atomic_read(&cache->stats.write_miss); atomic_read 1672 drivers/md/dm-cache-target.c sector_t current_volume = (atomic_read(&cache->nr_io_migrations) + 1) * atomic_read 2868 drivers/md/dm-cache-target.c BUG_ON(atomic_read(&cache->nr_io_migrations)); atomic_read 3198 drivers/md/dm-cache-target.c (unsigned) atomic_read(&cache->stats.read_hit), atomic_read 3199 drivers/md/dm-cache-target.c (unsigned) atomic_read(&cache->stats.read_miss), atomic_read 3200 drivers/md/dm-cache-target.c (unsigned) atomic_read(&cache->stats.write_hit), atomic_read 3201 drivers/md/dm-cache-target.c (unsigned) atomic_read(&cache->stats.write_miss), atomic_read 3202 drivers/md/dm-cache-target.c (unsigned) atomic_read(&cache->stats.demotion), atomic_read 3203 drivers/md/dm-cache-target.c (unsigned) atomic_read(&cache->stats.promotion), atomic_read 3204 drivers/md/dm-cache-target.c (unsigned long) atomic_read(&cache->nr_dirty)); atomic_read 790 drivers/md/dm-clone-target.c !atomic_read(&clone->ios_in_flight)) atomic_read 1097 drivers/md/dm-clone-target.c !atomic_read(&clone->ios_in_flight) && atomic_read 1100 drivers/md/dm-clone-target.c current_volume = atomic_read(&clone->hydrations_in_flight); atomic_read 1479 drivers/md/dm-clone-target.c atomic_read(&clone->hydrations_in_flight)); atomic_read 2024 drivers/md/dm-clone-target.c wait_event(clone->hydration_stopped, !atomic_read(&clone->hydrations_in_flight)); atomic_read 248 drivers/md/dm-delay.c if (!c->delay || !atomic_read(&dc->may_delay)) atomic_read 1204 drivers/md/dm-era-target.c if (!atomic_read(&era->suspended)) atomic_read 777 drivers/md/dm-integrity.c BUG_ON(!atomic_read(&comp->in_flight)); atomic_read 1255 drivers/md/dm-ioctl.c priv->global_event_nr = atomic_read(&dm_global_event_nr); atomic_read 1942 drivers/md/dm-ioctl.c priv->global_event_nr = atomic_read(&dm_global_event_nr); atomic_read 1960 drivers/md/dm-ioctl.c if ((int)(atomic_read(&dm_global_event_nr) - priv->global_event_nr) > 0) atomic_read 973 drivers/md/dm-kcopyd.c wait_event(kc->destroyq, !atomic_read(&kc->nr_jobs)); atomic_read 332 drivers/md/dm-log-userspace-base.c if (atomic_read(&lc->sched_flush)) atomic_read 371 drivers/md/dm-log-userspace-base.c if (lc->integrated_flush && atomic_read(&lc->sched_flush)) atomic_read 605 drivers/md/dm-log-userspace-base.c if (mark_list_is_empty && !atomic_read(&lc->sched_flush)) { atomic_read 647 drivers/md/dm-log-writes.c wait_event(lc->wait, !atomic_read(&lc->io_blocks) && atomic_read 648 drivers/md/dm-log-writes.c !atomic_read(&lc->pending_blocks)); atomic_read 291 drivers/md/dm-mpath.c if (atomic_read(&m->pg_init_in_progress) || test_bit(MPATHF_PG_INIT_DISABLED, &m->flags)) atomic_read 312 drivers/md/dm-mpath.c return atomic_read(&m->pg_init_in_progress); atomic_read 375 drivers/md/dm-mpath.c if (!atomic_read(&m->nr_valid_paths)) { atomic_read 1135 drivers/md/dm-mpath.c unsigned nr_valid_paths = atomic_read(&m->nr_valid_paths); atomic_read 1182 drivers/md/dm-mpath.c if (!atomic_read(&m->pg_init_in_progress)) atomic_read 1196 drivers/md/dm-mpath.c if (atomic_read(&m->pg_init_in_progress)) atomic_read 1242 drivers/md/dm-mpath.c pgpath->path.dev->name, atomic_read(&m->nr_valid_paths)); atomic_read 1405 drivers/md/dm-mpath.c if (atomic_read(&m->pg_init_count) <= m->pg_init_retries && atomic_read 1550 drivers/md/dm-mpath.c if (atomic_read(&m->nr_valid_paths) == 0 && atomic_read 1584 drivers/md/dm-mpath.c if (atomic_read(&m->nr_valid_paths) == 0 && atomic_read 1681 drivers/md/dm-mpath.c atomic_read(&m->pg_init_count)); atomic_read 1940 drivers/md/dm-mpath.c if (atomic_read(&m->pg_init_in_progress)) atomic_read 1944 drivers/md/dm-mpath.c if (!atomic_read(&m->nr_valid_paths) && test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) atomic_read 100 drivers/md/dm-queue-length.c DMEMIT("%d ", atomic_read(&pi->qlen)); atomic_read 200 drivers/md/dm-queue-length.c (atomic_read(&pi->qlen) < atomic_read(&best->qlen))) atomic_read 203 drivers/md/dm-queue-length.c if (!atomic_read(&best->qlen)) atomic_read 175 drivers/md/dm-raid1.c return &ms->mirror[atomic_read(&ms->default_mirror)]; atomic_read 191 drivers/md/dm-raid1.c if (!atomic_read(&m->error_count)) atomic_read 420 drivers/md/dm-raid1.c if (likely(!atomic_read(&m->error_count))) atomic_read 434 drivers/md/dm-raid1.c return !atomic_read(&default_mirror->error_count); atomic_read 480 drivers/md/dm-raid1.c if (atomic_read(&ms->suspend)) { atomic_read 574 drivers/md/dm-raid1.c else if (m && atomic_read(&m->error_count)) atomic_read 1383 drivers/md/dm-raid1.c if (!atomic_read(&(m->error_count))) atomic_read 246 drivers/md/dm-region-hash.c BUG_ON(atomic_read(®->pending)); atomic_read 612 drivers/md/dm-region-hash.c if (atomic_read(®->pending)) atomic_read 680 drivers/md/dm-region-hash.c return atomic_read(&rh->recovery_in_flight); atomic_read 95 drivers/md/dm-service-time.c DMEMIT("%d %u ", atomic_read(&pi->in_flight_size), atomic_read 217 drivers/md/dm-service-time.c sz1 = atomic_read(&pi1->in_flight_size); atomic_read 218 drivers/md/dm-service-time.c sz2 = atomic_read(&pi2->in_flight_size); atomic_read 1488 drivers/md/dm-snap.c while (atomic_read(&s->pending_exceptions_count)) atomic_read 187 drivers/md/dm-stats.c return atomic_read(&shared->in_flight[READ]) + atomic_read 188 drivers/md/dm-stats.c atomic_read(&shared->in_flight[WRITE]); atomic_read 224 drivers/md/dm-stats.c atomic_read(&shared->in_flight[READ]), atomic_read 225 drivers/md/dm-stats.c atomic_read(&shared->in_flight[WRITE])); atomic_read 501 drivers/md/dm-stats.c in_flight_read = (unsigned)atomic_read(&shared->in_flight[READ]); atomic_read 502 drivers/md/dm-stats.c in_flight_write = (unsigned)atomic_read(&shared->in_flight[WRITE]); atomic_read 410 drivers/md/dm-stripe.c DMEMIT("%c", atomic_read(&(sc->stripe[i].error_count)) ? atomic_read 453 drivers/md/dm-stripe.c if (atomic_read(&(sc->stripe[i].error_count)) < atomic_read 440 drivers/md/dm-writecache.c BUG_ON(atomic_read(&endio->count) <= 0); atomic_read 448 drivers/md/dm-writecache.c !atomic_read(&wc->bio_in_progress[direction])); atomic_read 217 drivers/md/dm-zoned-metadata.c return atomic_read(&zmd->unmap_nr_rnd); atomic_read 275 drivers/md/dm-zoned-metadata.c if (zmd->max_nr_mblks && atomic_read(&zmd->nr_mblks) > zmd->max_nr_mblks) { atomic_read 466 drivers/md/dm-zoned-metadata.c atomic_read(&zmd->nr_mblks) > zmd->min_nr_mblks && atomic_read 487 drivers/md/dm-zoned-metadata.c return atomic_read(&zmd->nr_mblks); atomic_read 1457 drivers/md/dm-zoned-metadata.c } else if (atomic_read(&zmd->nr_reserved_seq_zones) < zmd->nr_reserved_seq) { atomic_read 1855 drivers/md/dm-zoned-metadata.c } else if (atomic_read(&zmd->nr_reserved_seq_zones) < atomic_read 2486 drivers/md/dm-zoned-metadata.c zmd->nr_rnd, atomic_read(&zmd->unmap_nr_rnd)); atomic_read 2488 drivers/md/dm-zoned-metadata.c zmd->nr_seq, atomic_read(&zmd->unmap_nr_seq)); atomic_read 218 drivers/md/dm-zoned.h return atomic_read(&zone->refcount); atomic_read 374 drivers/md/dm.c return atomic_read(&md->open_count); atomic_read 2261 drivers/md/dm.c BUG_ON(!atomic_read(&md->holders)); atomic_read 2406 drivers/md/dm.c while (atomic_read(&md->holders)) atomic_read 2408 drivers/md/dm.c else if (atomic_read(&md->holders)) atomic_read 2410 drivers/md/dm.c dm_device_name(md), atomic_read(&md->holders)); atomic_read 2909 drivers/md/dm.c return atomic_read(&md->event_nr); atomic_read 2915 drivers/md/dm.c (event_nr != atomic_read(&md->event_nr))); atomic_read 311 drivers/md/md-bitmap.c atomic_read(&bitmap->pending_writes)==0); atomic_read 408 drivers/md/md-bitmap.c atomic_read(&bitmap->pending_writes)==0); atomic_read 433 drivers/md/md-bitmap.c atomic_read(&bitmap->pending_writes)==0); atomic_read 1404 drivers/md/md-bitmap.c bw = atomic_read(&bitmap->behind_writes); atomic_read 1469 drivers/md/md-bitmap.c atomic_read(&bitmap->behind_writes), atomic_read 1628 drivers/md/md-bitmap.c atomic_read(&bitmap->mddev->recovery_active) == 0); atomic_read 1752 drivers/md/md-bitmap.c atomic_read(&bitmap->pending_writes) == 0); atomic_read 1776 drivers/md/md-bitmap.c if (bitmap && atomic_read(&bitmap->behind_writes) > 0) { atomic_read 1781 drivers/md/md-bitmap.c atomic_read(&bitmap->behind_writes) == 0); atomic_read 88 drivers/md/md-faulty.c atomic_read(&conf->counters[mode]) <= 0) atomic_read 171 drivers/md/md-faulty.c if (atomic_read(&conf->counters[WriteAll])) { atomic_read 226 drivers/md/md-faulty.c if ((n=atomic_read(&conf->counters[WriteTransient])) != 0) atomic_read 230 drivers/md/md-faulty.c if ((n=atomic_read(&conf->counters[ReadTransient])) != 0) atomic_read 234 drivers/md/md-faulty.c if ((n=atomic_read(&conf->counters[WritePersistent])) != 0) atomic_read 238 drivers/md/md-faulty.c if ((n=atomic_read(&conf->counters[ReadPersistent])) != 0) atomic_read 243 drivers/md/md-faulty.c if ((n=atomic_read(&conf->counters[ReadFixable])) != 0) atomic_read 247 drivers/md/md-faulty.c if ((n=atomic_read(&conf->counters[WriteAll])) != 0) atomic_read 282 drivers/md/md-multipath.c atomic_read(&rdev->nr_pending)) { atomic_read 290 drivers/md/md-multipath.c if (atomic_read(&rdev->nr_pending)) { atomic_read 431 drivers/md/md.c wait_event(mddev->sb_wait, atomic_read(&mddev->active_io) == 0); atomic_read 910 drivers/md/md.c wait_event(mddev->sb_wait, atomic_read(&mddev->pending_writes)==0); atomic_read 1923 drivers/md/md.c sb->cnt_corrected_read = cpu_to_le32(atomic_read(&rdev->corrected_errors)); atomic_read 3012 drivers/md/md.c return sprintf(page, "%d\n", atomic_read(&rdev->corrected_errors)); atomic_read 4398 drivers/md/md.c atomic_read(&mddev->max_corr_read_errors)); atomic_read 4865 drivers/md/md.c resync = mddev->curr_mark_cnt - atomic_read(&mddev->recovery_active); atomic_read 6109 drivers/md/md.c if ((mddev->pers && atomic_read(&mddev->openers) > !!bdev) || atomic_read 6170 drivers/md/md.c if ((mddev->pers && atomic_read(&mddev->openers) > !!bdev) || atomic_read 6848 drivers/md/md.c } else if (atomic_read(&inode->i_writecount) != 1) { atomic_read 7368 drivers/md/md.c if (mddev->pers && atomic_read(&mddev->openers) > 1) { atomic_read 7833 drivers/md/md.c resync -= atomic_read(&mddev->recovery_active); atomic_read 7921 drivers/md/md.c recovery_active = atomic_read(&mddev->recovery_active); atomic_read 8016 drivers/md/md.c seq->poll_event = atomic_read(&md_event_count); atomic_read 8116 drivers/md/md.c seq->poll_event = atomic_read(&md_event_count); atomic_read 8133 drivers/md/md.c if (seq->poll_event != atomic_read(&md_event_count)) atomic_read 8228 drivers/md/md.c atomic_read(&disk->sync_io); atomic_read 8628 drivers/md/md.c atomic_read(&mddev->recovery_active) == 0); atomic_read 8693 drivers/md/md.c mark_cnt[next] = io_sectors - atomic_read(&mddev->recovery_active); atomic_read 8710 drivers/md/md.c recovery_done = io_sectors - atomic_read(&mddev->recovery_active); atomic_read 8725 drivers/md/md.c !atomic_read(&mddev->recovery_active)); atomic_read 8736 drivers/md/md.c wait_event(mddev->recovery_wait, !atomic_read(&mddev->recovery_active)); atomic_read 8838 drivers/md/md.c atomic_read(&rdev->nr_pending)==0) { atomic_read 8858 drivers/md/md.c atomic_read(&rdev->nr_pending)==0)) { atomic_read 518 drivers/md/raid1.c if (atomic_read(&r1_bio->behind_remaining) >= (atomic_read(&r1_bio->remaining)-1) && atomic_read 690 drivers/md/raid1.c pending = atomic_read(&rdev->nr_pending); atomic_read 889 drivers/md/raid1.c !atomic_read(&conf->nr_waiting[idx]), atomic_read 913 drivers/md/raid1.c !atomic_read(&conf->nr_pending[idx]) && atomic_read 914 drivers/md/raid1.c atomic_read(&conf->barrier[idx]) < RESYNC_DEPTH) || atomic_read 935 drivers/md/raid1.c BUG_ON(atomic_read(&conf->barrier[idx]) <= 0); atomic_read 973 drivers/md/raid1.c !atomic_read(&conf->barrier[idx])) atomic_read 994 drivers/md/raid1.c !atomic_read(&conf->barrier[idx]), atomic_read 1059 drivers/md/raid1.c ret = atomic_read(&conf->nr_sync_pending); atomic_read 1061 drivers/md/raid1.c ret += atomic_read(&conf->nr_pending[idx]) - atomic_read 1062 drivers/md/raid1.c atomic_read(&conf->nr_queued[idx]); atomic_read 1290 drivers/md/raid1.c atomic_read(&bitmap->behind_writes) == 0); atomic_read 1490 drivers/md/raid1.c (atomic_read(&bitmap->behind_writes) atomic_read 1819 drivers/md/raid1.c atomic_read(&rdev->nr_pending)) { atomic_read 1835 drivers/md/raid1.c if (atomic_read(&rdev->nr_pending)) { atomic_read 1850 drivers/md/raid1.c if (atomic_read(&repl->nr_pending)) { atomic_read 2697 drivers/md/raid1.c if (atomic_read(&conf->nr_waiting[idx])) atomic_read 800 drivers/md/raid10.c pending = atomic_read(&rdev->nr_pending); atomic_read 964 drivers/md/raid10.c !atomic_read(&conf->nr_pending) && conf->barrier < RESYNC_DEPTH, atomic_read 996 drivers/md/raid10.c (atomic_read(&conf->nr_pending) && atomic_read 1035 drivers/md/raid10.c atomic_read(&conf->nr_pending) == conf->nr_queued+extra, atomic_read 1839 drivers/md/raid10.c atomic_read(&rdev->nr_pending)) { atomic_read 1857 drivers/md/raid10.c if (atomic_read(&rdev->nr_pending)) { atomic_read 2272 drivers/md/raid10.c unsigned int read_errors = atomic_read(&rdev->read_errors); atomic_read 2335 drivers/md/raid10.c int max_read_errors = atomic_read(&mddev->max_corr_read_errors); atomic_read 2350 drivers/md/raid10.c if (atomic_read(&rdev->read_errors) > max_read_errors) { atomic_read 2356 drivers/md/raid10.c atomic_read(&rdev->read_errors), max_read_errors); atomic_read 336 drivers/md/raid5-cache.c total_cached = atomic_read(&conf->r5c_cached_partial_stripes) + atomic_read 337 drivers/md/raid5-cache.c atomic_read(&conf->r5c_cached_full_stripes); atomic_read 348 drivers/md/raid5-cache.c atomic_read(&conf->empty_inactive_list_nr) > 0) atomic_read 365 drivers/md/raid5-cache.c if (atomic_read(&conf->r5c_cached_full_stripes) >= atomic_read 407 drivers/md/raid5-cache.c ((conf->max_degraded + 1) * atomic_read(&log->stripe_in_journal_count) + atomic_read 1437 drivers/md/raid5-cache.c flushing_partial = atomic_read(&conf->r5c_flushing_partial_stripes); atomic_read 1438 drivers/md/raid5-cache.c flushing_full = atomic_read(&conf->r5c_flushing_full_stripes); atomic_read 1439 drivers/md/raid5-cache.c total_cached = atomic_read(&conf->r5c_cached_partial_stripes) + atomic_read 1440 drivers/md/raid5-cache.c atomic_read(&conf->r5c_cached_full_stripes) - atomic_read 1444 drivers/md/raid5-cache.c atomic_read(&conf->empty_inactive_list_nr) > 0) atomic_read 1451 drivers/md/raid5-cache.c atomic_read(&conf->r5c_cached_full_stripes) - flushing_full > atomic_read 1483 drivers/md/raid5-cache.c atomic_read(&sh->count) == 0) { atomic_read 2448 drivers/md/raid5-cache.c atomic_read(&conf->active_stripes) == 0); atomic_read 2870 drivers/md/raid5-cache.c BUG_ON(atomic_read(&conf->r5c_cached_partial_stripes) == 0); atomic_read 2876 drivers/md/raid5-cache.c BUG_ON(atomic_read(&conf->r5c_cached_full_stripes) == 0); atomic_read 219 drivers/md/raid5.c BUG_ON(atomic_read(&conf->active_stripes)==0); atomic_read 346 drivers/md/raid5.c if (atomic_read(&conf->active_stripes) == 0) atomic_read 493 drivers/md/raid5.c BUG_ON(atomic_read(&sh->count) != 0); atomic_read 661 drivers/md/raid5.c (atomic_read(&conf->active_stripes) atomic_read 674 drivers/md/raid5.c if (!atomic_read(&sh->count)) { atomic_read 758 drivers/md/raid5.c if (!atomic_read(&head->count)) { atomic_read 2443 drivers/md/raid5.c BUG_ON(atomic_read(&sh->count)); atomic_read 2475 drivers/md/raid5.c (unsigned long long)sh->sector, i, atomic_read(&sh->count), atomic_read 2521 drivers/md/raid5.c if (atomic_read(&rdev->read_errors)) atomic_read 2552 drivers/md/raid5.c } else if (atomic_read(&rdev->read_errors) atomic_read 2557 drivers/md/raid5.c atomic_read(&rdev->read_errors), atomic_read 2621 drivers/md/raid5.c (unsigned long long)sh->sector, i, atomic_read(&sh->count), atomic_read 4714 drivers/md/raid5.c atomic_read(&sh->count), sh->pd_idx, sh->qd_idx, atomic_read 5053 drivers/md/raid5.c if (atomic_read(&conf->preread_active_stripes) < atomic_read 5063 drivers/md/raid5.c if (atomic_read(&conf->preread_active_stripes) < IO_THRESHOLD) { atomic_read 5111 drivers/md/raid5.c if (atomic_read(&conf->empty_inactive_list_nr)) atomic_read 5360 drivers/md/raid5.c atomic_read(&conf->pending_full_writes), conf->bypass_count); atomic_read 5380 drivers/md/raid5.c atomic_read(&conf->pending_full_writes) == 0)) { atomic_read 5883 drivers/md/raid5.c atomic_read(&conf->reshape_stripes)==0 atomic_read 5885 drivers/md/raid5.c if (atomic_read(&conf->reshape_stripes) != 0) atomic_read 5990 drivers/md/raid5.c atomic_read(&conf->reshape_stripes) == 0 atomic_read 5992 drivers/md/raid5.c if (atomic_read(&conf->reshape_stripes) != 0) atomic_read 6578 drivers/md/raid5.c return sprintf(page, "%d\n", atomic_read(&conf->active_stripes)); atomic_read 7611 drivers/md/raid5.c if (atomic_read(&conf->active_stripes) || atomic_read 7612 drivers/md/raid5.c atomic_read(&conf->r5c_cached_full_stripes) || atomic_read 7613 drivers/md/raid5.c atomic_read(&conf->r5c_cached_partial_stripes)) { atomic_read 7631 drivers/md/raid5.c atomic_read(&rdev->nr_pending)) { atomic_read 7649 drivers/md/raid5.c if (atomic_read(&rdev->nr_pending)) { atomic_read 8082 drivers/md/raid5.c atomic_read(&conf->active_stripes) == 0 && atomic_read 8083 drivers/md/raid5.c atomic_read(&conf->active_aligned_reads) == 0, atomic_read 117 drivers/media/cec/cec-pin.c if (atomic_read(&pin->work_pin_num_events) < CEC_NUM_PIN_EVENTS) { atomic_read 1042 drivers/media/cec/cec-pin.c atomic_read(&pin->work_irq_change) || atomic_read 1043 drivers/media/cec/cec-pin.c atomic_read(&pin->work_pin_num_events)); atomic_read 1071 drivers/media/cec/cec-pin.c while (atomic_read(&pin->work_pin_num_events)) { atomic_read 1474 drivers/media/common/videobuf2/videobuf2-core.c atomic_read(&q->owned_by_drv_count)); atomic_read 1487 drivers/media/common/videobuf2/videobuf2-core.c if (WARN_ON(atomic_read(&q->owned_by_drv_count))) { atomic_read 1500 drivers/media/common/videobuf2/videobuf2-core.c WARN_ON(atomic_read(&q->owned_by_drv_count)); atomic_read 1781 drivers/media/common/videobuf2/videobuf2-core.c wait_event(q->done_wq, !atomic_read(&q->owned_by_drv_count)); atomic_read 1882 drivers/media/common/videobuf2/videobuf2-core.c if (WARN_ON(atomic_read(&q->owned_by_drv_count))) { atomic_read 1890 drivers/media/common/videobuf2/videobuf2-core.c WARN_ON(atomic_read(&q->owned_by_drv_count)); atomic_read 231 drivers/media/dvb-core/dvb_ca_en50221.c return (atomic_read(&sl->camchange_count) != 0); atomic_read 68 drivers/media/dvb-frontends/cxd2880/cxd2880_integ.c if (atomic_read(&tnr_dmd->cancel) != 0) atomic_read 661 drivers/media/i2c/cx25840/cx25840-ir.c invert = (bool) atomic_read(&ir_state->rx_invert); atomic_read 662 drivers/media/i2c/cx25840/cx25840-ir.c divider = (u16) atomic_read(&ir_state->rxclk_divider); atomic_read 24 drivers/media/pci/cx18/cx18-controls.c if (atomic_read(&cx->ana_capturing) > 0) atomic_read 1262 drivers/media/pci/cx18/cx18-driver.c if (atomic_read(&cx->tot_capturing) > 0) atomic_read 232 drivers/media/pci/cx18/cx18-fileops.c if (!atomic_read(&s->q_full.depth)) atomic_read 412 drivers/media/pci/cx18/cx18-fileops.c if (atomic_read(&cx->ana_capturing) == 0 && s->id == -1) { atomic_read 639 drivers/media/pci/cx18/cx18-fileops.c if (atomic_read(&s->q_full.depth)) atomic_read 759 drivers/media/pci/cx18/cx18-fileops.c if (atomic_read(&cx->ana_capturing) > 0) { atomic_read 805 drivers/media/pci/cx18/cx18-fileops.c if (atomic_read(&cx->ana_capturing) > 0) { atomic_read 851 drivers/media/pci/cx18/cx18-fileops.c if (atomic_read(&cx->ana_capturing)) { atomic_read 864 drivers/media/pci/cx18/cx18-fileops.c if (atomic_read(&cx->ana_capturing)) { atomic_read 273 drivers/media/pci/cx18/cx18-ioctl.c if (atomic_read(&cx->ana_capturing) > 0) atomic_read 305 drivers/media/pci/cx18/cx18-ioctl.c if (!cx18_raw_vbi(cx) && atomic_read(&cx->ana_capturing) > 0) atomic_read 338 drivers/media/pci/cx18/cx18-ioctl.c if (cx18_raw_vbi(cx) && atomic_read(&cx->ana_capturing) > 0) atomic_read 578 drivers/media/pci/cx18/cx18-ioctl.c atomic_read(&cx->ana_capturing) > 0) { atomic_read 932 drivers/media/pci/cx18/cx18-ioctl.c if (!atomic_read(&cx->ana_capturing)) atomic_read 948 drivers/media/pci/cx18/cx18-ioctl.c if (!atomic_read(&cx->ana_capturing)) atomic_read 1034 drivers/media/pci/cx18/cx18-ioctl.c atomic_read(&s->q_full.depth) * s->bufs_per_mdl * 100 atomic_read 511 drivers/media/pci/cx18/cx18-mailbox.c if (atomic_read(&cx->in_work_order[i].pending) == 0) { atomic_read 57 drivers/media/pci/cx18/cx18-queue.c atomic_read(&q->depth) >= CX18_MAX_FW_MDLS_PER_STREAM) atomic_read 151 drivers/media/pci/cx18/cx18-queue.c if (mdl->skipped >= atomic_read(&s->q_busy.depth)-1) { atomic_read 656 drivers/media/pci/cx18/cx18-streams.c if ((atomic_read(&s->q_free.depth) + atomic_read(&s->q_busy.depth)) >= atomic_read 661 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&s->q_full.depth) < 2) atomic_read 703 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&s->q_free.depth) == 0 || atomic_read 704 drivers/media/pci/cx18/cx18-streams.c atomic_read(&s->q_busy.depth) >= CX18_MAX_FW_MDLS_PER_STREAM) atomic_read 713 drivers/media/pci/cx18/cx18-streams.c } while (atomic_read(&s->q_busy.depth) < CX18_MAX_FW_MDLS_PER_STREAM atomic_read 850 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&cx->ana_capturing) == 0) atomic_read 902 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&cx->tot_capturing) == 0) { atomic_read 931 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&cx->tot_capturing) == 0) { atomic_read 972 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&cx->tot_capturing) == 0) atomic_read 999 drivers/media/pci/cx18/cx18-streams.c if (atomic_read(&cx->tot_capturing) > 0) atomic_read 477 drivers/media/pci/cx23885/cx23885-alsa.c count = atomic_read(&chip->count); atomic_read 109 drivers/media/pci/cx23885/cx23885-input.c params.shutdown = atomic_read(&dev->ir_input_stopping); atomic_read 120 drivers/media/pci/cx23885/cx23885-input.c params.shutdown = atomic_read(&dev->ir_input_stopping); atomic_read 653 drivers/media/pci/cx23885/cx23888-ir.c bool invert = (bool) atomic_read(&state->rx_invert); atomic_read 654 drivers/media/pci/cx23885/cx23888-ir.c u16 divider = (u16) atomic_read(&state->rxclk_divider); atomic_read 620 drivers/media/pci/cx25821/cx25821-alsa.c count = atomic_read(&chip->count); atomic_read 563 drivers/media/pci/cx88/cx88-alsa.c count = atomic_read(&chip->count); atomic_read 576 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->vbb.sequence = atomic_read(&q->frame_sequence); atomic_read 604 drivers/media/pci/intel/ipu3/ipu3-cio2.c .u.frame_sync.frame_sequence = atomic_read(&q->frame_sequence), atomic_read 106 drivers/media/pci/ivtv/ivtv-controls.c if (atomic_read(&itv->decoding)) { atomic_read 1405 drivers/media/pci/ivtv/ivtv-driver.c if (atomic_read(&itv->capturing) > 0) atomic_read 1414 drivers/media/pci/ivtv/ivtv-driver.c if (atomic_read(&itv->decoding) > 0) { atomic_read 352 drivers/media/pci/ivtv/ivtv-fileops.c if (atomic_read(&itv->capturing) == 0 && s->fh == NULL) { atomic_read 517 drivers/media/pci/ivtv/ivtv-fileops.c if (atomic_read(&itv->decoding) == 0) { atomic_read 892 drivers/media/pci/ivtv/ivtv-fileops.c if (atomic_read(&itv->capturing) > 0) { atomic_read 996 drivers/media/pci/ivtv/ivtv-fileops.c if (atomic_read(&itv->capturing) > 0) { atomic_read 1048 drivers/media/pci/ivtv/ivtv-fileops.c if (atomic_read(&itv->capturing)) atomic_read 1055 drivers/media/pci/ivtv/ivtv-fileops.c if (atomic_read(&itv->capturing)) { atomic_read 334 drivers/media/pci/ivtv/ivtv-firmware.c if (!res && !atomic_read(&itv->capturing) && atomic_read 335 drivers/media/pci/ivtv/ivtv-firmware.c (!atomic_read(&itv->decoding) || atomic_read 336 drivers/media/pci/ivtv/ivtv-firmware.c (atomic_read(&itv->decoding) < 2 && test_bit(IVTV_F_I_DEC_YUV, atomic_read 365 drivers/media/pci/ivtv/ivtv-firmware.c if (res && !atomic_read(&itv->capturing) && atomic_read 366 drivers/media/pci/ivtv/ivtv-firmware.c !atomic_read(&itv->decoding)) { atomic_read 163 drivers/media/pci/ivtv/ivtv-ioctl.c if (atomic_read(&itv->decoding) > 0) { atomic_read 272 drivers/media/pci/ivtv/ivtv-ioctl.c if (atomic_read(&itv->decoding) == 0) atomic_read 283 drivers/media/pci/ivtv/ivtv-ioctl.c if (!atomic_read(&itv->decoding)) atomic_read 287 drivers/media/pci/ivtv/ivtv-ioctl.c if (atomic_read(&itv->decoding) > 0) { atomic_read 297 drivers/media/pci/ivtv/ivtv-ioctl.c if (!atomic_read(&itv->decoding)) atomic_read 589 drivers/media/pci/ivtv/ivtv-ioctl.c if (atomic_read(&itv->capturing) > 0) atomic_read 607 drivers/media/pci/ivtv/ivtv-ioctl.c if (!ivtv_raw_vbi(itv) && atomic_read(&itv->capturing) > 0) atomic_read 626 drivers/media/pci/ivtv/ivtv-ioctl.c if (ivtv_raw_vbi(itv) && atomic_read(&itv->capturing) > 0) atomic_read 995 drivers/media/pci/ivtv/ivtv-ioctl.c if (atomic_read(&itv->capturing) > 0) { atomic_read 1173 drivers/media/pci/ivtv/ivtv-ioctl.c atomic_read(&itv->capturing) > 0 || atomic_read 1174 drivers/media/pci/ivtv/ivtv-ioctl.c atomic_read(&itv->decoding) > 0) { atomic_read 1267 drivers/media/pci/ivtv/ivtv-ioctl.c if (!atomic_read(&itv->capturing)) atomic_read 1302 drivers/media/pci/ivtv/ivtv-ioctl.c if (!atomic_read(&itv->capturing)) atomic_read 1315 drivers/media/pci/ivtv/ivtv-ioctl.c if (!atomic_read(&itv->capturing)) atomic_read 798 drivers/media/pci/ivtv/ivtv-irq.c if (atomic_read(&itv->yuv_info.next_dma_frame) >= 0) atomic_read 833 drivers/media/pci/ivtv/ivtv-irq.c int last_dma_frame = atomic_read(&yi->next_dma_frame); atomic_read 844 drivers/media/pci/ivtv/ivtv-irq.c if (next_dma_frame >= 0 && next_dma_frame != atomic_read(&yi->next_fill_frame)) { atomic_read 898 drivers/media/pci/ivtv/ivtv-irq.c (u8)(atomic_read(&yi->next_dma_frame) - atomic_read 549 drivers/media/pci/ivtv/ivtv-streams.c if (atomic_read(&itv->capturing) == 0) { atomic_read 617 drivers/media/pci/ivtv/ivtv-streams.c if (atomic_read(&itv->capturing) == 0) { atomic_read 808 drivers/media/pci/ivtv/ivtv-streams.c if (atomic_read(&itv->capturing) == 0) atomic_read 887 drivers/media/pci/ivtv/ivtv-streams.c if (atomic_read(&itv->capturing) > 0) { atomic_read 1013 drivers/media/pci/ivtv/ivtv-streams.c if (atomic_read(&itv->capturing) == 0) { atomic_read 1036 drivers/media/pci/ivtv/ivtv-streams.c if (atomic_read(&itv->capturing) == 0) atomic_read 947 drivers/media/pci/ivtv/ivtv-yuv.c if (atomic_read(&yi->next_dma_frame) == -1) atomic_read 950 drivers/media/pci/ivtv/ivtv-yuv.c draw = atomic_read(&yi->next_fill_frame); atomic_read 951 drivers/media/pci/ivtv/ivtv-yuv.c display = atomic_read(&yi->next_dma_frame); atomic_read 66 drivers/media/platform/exynos4-is/fimc-lite-reg.c if (atomic_read(&dev->out_path) == FIMC_IO_DMA) { atomic_read 270 drivers/media/platform/exynos4-is/fimc-lite.c if (atomic_read(&fimc->out_path) != FIMC_IO_DMA) atomic_read 465 drivers/media/platform/exynos4-is/fimc-lite.c if (atomic_read(&fimc->out_path) != FIMC_IO_DMA) { atomic_read 480 drivers/media/platform/exynos4-is/fimc-lite.c atomic_read(&fimc->out_path) != FIMC_IO_DMA) atomic_read 515 drivers/media/platform/exynos4-is/fimc-lite.c atomic_read(&fimc->out_path) == FIMC_IO_DMA) { atomic_read 1084 drivers/media/platform/exynos4-is/fimc-lite.c if ((atomic_read(&fimc->out_path) == FIMC_IO_ISP && atomic_read 1086 drivers/media/platform/exynos4-is/fimc-lite.c (atomic_read(&fimc->out_path) == FIMC_IO_DMA && atomic_read 1213 drivers/media/platform/exynos4-is/fimc-lite.c if (atomic_read(&fimc->out_path) != FIMC_IO_ISP) atomic_read 1583 drivers/media/platform/exynos4-is/fimc-lite.c fimc_lite_hw_init(fimc, atomic_read(&fimc->out_path) == FIMC_IO_ISP); atomic_read 155 drivers/media/platform/m2m-deinterlace.c !atomic_read(&ctx->dev->busy)) { atomic_read 1222 drivers/media/platform/omap3isp/isp.c if (!wait_event_timeout(*wait, !atomic_read(stopping), atomic_read 1467 drivers/media/platform/omap3isp/ispccdc.c event.u.frame_sync.frame_sequence = atomic_read(&pipe->frame_number); atomic_read 275 drivers/media/platform/omap3isp/isphist.c if (atomic_read(&hist->buf_err) || hist->state != ISPSTAT_ENABLED) { atomic_read 918 drivers/media/platform/omap3isp/ispstat.c stat->frame_number = atomic_read(&pipe->frame_number); atomic_read 544 drivers/media/platform/omap3isp/ispvideo.c buf->vb.sequence = atomic_read(&pipe->frame_number); atomic_read 978 drivers/media/platform/qcom/camss/camss.c if (atomic_read(&camss->ref_count) == 0) atomic_read 103 drivers/media/platform/qcom/venus/hfi.c !atomic_read(&core->insts_count)); atomic_read 150 drivers/media/platform/s5p-mfc/s5p_mfc.c if (atomic_read(&dev->watchdog_cnt) >= MFC_WATCHDOG_CNT) { atomic_read 64 drivers/media/platform/s5p-mfc/s5p_mfc_pm.c mfc_debug(3, "+ %d\n", atomic_read(&clk_ref)); atomic_read 72 drivers/media/platform/s5p-mfc/s5p_mfc_pm.c mfc_debug(3, "- %d\n", atomic_read(&clk_ref)); atomic_read 170 drivers/media/platform/sti/c8sectpfe/c8sectpfe-core.c if (!atomic_read(&fei->fw_loaded)) { atomic_read 267 drivers/media/platform/sti/c8sectpfe/c8sectpfe-core.c if (!atomic_read(&fei->fw_loaded)) { atomic_read 165 drivers/media/radio/radio-shark.c brightness = atomic_read(&shark->brightness[i]); atomic_read 152 drivers/media/radio/radio-shark2.c brightness = atomic_read(&shark->brightness[i]); atomic_read 1084 drivers/media/radio/radio-si476x.c atomic_read(&radio->core->is_alive)) atomic_read 1109 drivers/media/radio/radio-si476x.c !atomic_read(&radio->core->is_alive))); atomic_read 1113 drivers/media/radio/radio-si476x.c if (!atomic_read(&radio->core->is_alive)) atomic_read 1140 drivers/media/radio/radio-si476x.c if (atomic_read(&radio->core->is_alive)) atomic_read 1143 drivers/media/radio/radio-si476x.c if (!atomic_read(&radio->core->is_alive)) atomic_read 324 drivers/media/radio/wl128x/fmdrv_common.c if (num_fm_hci_cmds && atomic_read(&fmdev->tx_cnt)) atomic_read 339 drivers/media/radio/wl128x/fmdrv_common.c if (!atomic_read(&fmdev->tx_cnt)) atomic_read 110 drivers/media/usb/cx231xx/cx231xx-audio.c if (atomic_read(&dev->stream_started) == 0) atomic_read 201 drivers/media/usb/cx231xx/cx231xx-audio.c if (atomic_read(&dev->stream_started) == 0) atomic_read 500 drivers/media/usb/cx231xx/cx231xx-audio.c if (atomic_read(&dev->stream_started) > 0) { atomic_read 538 drivers/media/usb/cx231xx/cx231xx-audio.c if (atomic_read(&dev->stream_started) > 0) { atomic_read 560 drivers/media/usb/cx231xx/cx231xx-audio.c if (atomic_read(&dev->stream_started)) { atomic_read 65 drivers/media/usb/cx231xx/cx231xx-core.c if (atomic_read(&dev->devlist_count) > 0) { atomic_read 1897 drivers/media/usb/cx231xx/cx231xx-video.c if (atomic_read(&dev->devlist_count) > 0) { atomic_read 110 drivers/media/usb/em28xx/em28xx-audio.c if (atomic_read(&dev->adev.stream_started) == 0) atomic_read 339 drivers/media/usb/em28xx/em28xx-audio.c if (atomic_read(&dev->adev.stream_started) > 0) { atomic_read 394 drivers/media/usb/em28xx/em28xx-audio.c if (atomic_read(&adev->stream_started) > 0) { atomic_read 421 drivers/media/usb/em28xx/em28xx-audio.c if (atomic_read(&adev->stream_started)) { atomic_read 1270 drivers/media/usb/gspca/cpia1.c framerate = atomic_read(&sd->fps); atomic_read 1393 drivers/media/usb/gspca/cpia1.c cam_exposure = atomic_read(&sd->cam_exposure); atomic_read 343 drivers/media/usb/gspca/pac207.c int avg_lum = atomic_read(&sd->avg_lum); atomic_read 727 drivers/media/usb/gspca/pac7302.c int avg_lum = atomic_read(&sd->avg_lum); atomic_read 492 drivers/media/usb/gspca/pac7311.c int avg_lum = atomic_read(&sd->avg_lum); atomic_read 2142 drivers/media/usb/gspca/sn9c20x.c avg_lum = atomic_read(&sd->avg_lum); atomic_read 880 drivers/media/usb/gspca/sonixb.c avg_lum = atomic_read(&sd->avg_lum); atomic_read 2584 drivers/media/usb/gspca/sonixj.c delta = atomic_read(&sd->avg_lum); atomic_read 1416 drivers/media/usb/s2255/s2255drv.c state = atomic_read(&dev->fw_data->fw_state); atomic_read 1425 drivers/media/usb/s2255/s2255drv.c ((atomic_read(&dev->fw_data->fw_state) atomic_read 1427 drivers/media/usb/s2255/s2255drv.c (atomic_read(&dev->fw_data->fw_state) atomic_read 1431 drivers/media/usb/s2255/s2255drv.c state = atomic_read(&dev->fw_data->fw_state); atomic_read 1439 drivers/media/usb/s2255/s2255drv.c ((atomic_read(&dev->fw_data->fw_state) atomic_read 1441 drivers/media/usb/s2255/s2255drv.c (atomic_read(&dev->fw_data->fw_state) atomic_read 1445 drivers/media/usb/s2255/s2255drv.c state = atomic_read(&dev->fw_data->fw_state); atomic_read 1555 drivers/media/usb/s2255/s2255drv.c atomic_read(&dev->num_channels)); atomic_read 1672 drivers/media/usb/s2255/s2255drv.c if (atomic_read(&dev->num_channels) == 0) { atomic_read 1676 drivers/media/usb/s2255/s2255drv.c if (atomic_read(&dev->num_channels) != MAX_CHANNELS) atomic_read 2345 drivers/media/usb/s2255/s2255drv.c int channels = atomic_read(&dev->num_channels); atomic_read 506 drivers/media/usb/stkwebcam/stk-webcam.c if (atomic_read(&dev->urbs_used) && is_present(dev)) atomic_read 193 drivers/media/usb/tm6000/tm6000-alsa.c if (atomic_read(&core->stream_started) > 0) { atomic_read 210 drivers/media/usb/tm6000/tm6000-alsa.c if (atomic_read(&core->stream_started) == 0) atomic_read 296 drivers/media/usb/tm6000/tm6000-alsa.c if (atomic_read(&core->stream_started) > 0) { atomic_read 328 drivers/media/usb/tm6000/tm6000-alsa.c if (atomic_read(&core->stream_started)) { atomic_read 80 drivers/media/usb/usbtv/usbtv-audio.c if (atomic_read(&chip->snd_stream)) { atomic_read 146 drivers/media/usb/usbtv/usbtv-audio.c if (!atomic_read(&chip->snd_stream)) atomic_read 282 drivers/media/usb/usbtv/usbtv-audio.c if (atomic_read(&usbtv->snd_stream) && usbtv->snd_bulk_urb) atomic_read 288 drivers/media/usb/usbtv/usbtv-audio.c if (atomic_read(&usbtv->snd_stream) && usbtv->snd_bulk_urb) atomic_read 299 drivers/media/usb/usbtv/usbtv-audio.c if (atomic_read(&chip->snd_stream)) atomic_read 270 drivers/media/v4l2-core/v4l2-clk.c if (WARN(atomic_read(&clk->use_count), atomic_read 287 drivers/media/v4l2-core/v4l2-dev.c if (atomic_read(&global->prios[V4L2_PRIORITY_RECORD]) > 0) atomic_read 289 drivers/media/v4l2-core/v4l2-dev.c if (atomic_read(&global->prios[V4L2_PRIORITY_INTERACTIVE]) > 0) atomic_read 291 drivers/media/v4l2-core/v4l2-dev.c if (atomic_read(&global->prios[V4L2_PRIORITY_BACKGROUND]) > 0) atomic_read 508 drivers/message/fusion/mptlan.c priv->total_posted,atomic_read(&priv->buckets_out))); atomic_read 515 drivers/message/fusion/mptlan.c while (atomic_read(&priv->buckets_out) && time_before(jiffies, timeout)) atomic_read 835 drivers/message/fusion/mptlan.c atomic_read(&priv->buckets_out))); atomic_read 837 drivers/message/fusion/mptlan.c if (atomic_read(&priv->buckets_out) < priv->bucketthresh) atomic_read 842 drivers/message/fusion/mptlan.c atomic_read(&priv->buckets_out), priv->total_received)); atomic_read 953 drivers/message/fusion/mptlan.c /**/ atomic_read(&priv->buckets_out), priv->total_received)); atomic_read 1110 drivers/message/fusion/mptlan.c atomic_read(&priv->buckets_out)); atomic_read 1115 drivers/message/fusion/mptlan.c remaining, atomic_read(&priv->buckets_out)); atomic_read 1118 drivers/message/fusion/mptlan.c ((atomic_read(&priv->buckets_out) - remaining) > atomic_read 1154 drivers/message/fusion/mptlan.c curr = atomic_read(&priv->buckets_out); atomic_read 1280 drivers/message/fusion/mptlan.c __func__, buckets, atomic_read(&priv->buckets_out))); atomic_read 3759 drivers/message/fusion/mptsas.c atomic_read(&sdev->device_busy))); atomic_read 606 drivers/mfd/ab8500-core.c if (atomic_read(&ab8500->transfer_ongoing)) atomic_read 2344 drivers/mfd/db8500-prcmu.c return (atomic_read(&ac_wake_req_state) != 0); atomic_read 57 drivers/mfd/mfd-core.c WARN_ON(atomic_read(cell->usage_count) < 0); atomic_read 299 drivers/mfd/si476x-cmd.c atomic_read(&core->cts), atomic_read 314 drivers/mfd/si476x-cmd.c atomic_read(&core->cts), atomic_read 385 drivers/mfd/si476x-cmd.c atomic_read(&core->stc)); atomic_read 521 drivers/mfd/si476x-i2c.c if (atomic_read(&core->is_alive)) atomic_read 128 drivers/mfd/twl6030-irq.c chained_wakeups = atomic_read(&pdata->wakeirqs); atomic_read 290 drivers/misc/cb710/core.c BUG_ON(atomic_read(&chip->slot_refs_count) != 0); atomic_read 303 drivers/misc/cb710/core.c BUG_ON(atomic_read(&chip->slot_refs_count) != 0); atomic_read 358 drivers/misc/cxl/file.c if (ctx->afu_driver_ops && atomic_read(&ctx->afu_driver_events)) atomic_read 474 drivers/misc/cxl/file.c if (ctx->afu_driver_ops && atomic_read(&ctx->afu_driver_events)) { atomic_read 1126 drivers/misc/cxl/pci.c if (atomic_read(&afu->configured_state) != -1) { atomic_read 1962 drivers/misc/cxl/pci.c atomic_read(&adapter->contexts_num)); atomic_read 100 drivers/misc/habanalabs/command_buffer.c if ((hdev->disabled) || ((atomic_read(&hdev->in_reset)) && atomic_read 220 drivers/misc/habanalabs/command_buffer.c atomic_read(&hdev->in_reset) ? "in_reset" : "disabled"); atomic_read 624 drivers/misc/habanalabs/command_submission.c atomic_read(&hdev->in_reset) ? "in_reset" : "disabled"); atomic_read 21 drivers/misc/habanalabs/device.c if ((hdev->disabled) || (atomic_read(&hdev->in_reset))) atomic_read 33 drivers/misc/habanalabs/device.c else if (atomic_read(&hdev->in_reset)) atomic_read 249 drivers/misc/habanalabs/habanalabs_ioctl.c atomic_read(&hdev->in_reset) ? "in_reset" : "disabled"); atomic_read 302 drivers/misc/habanalabs/habanalabs_ioctl.c atomic_read(&hdev->in_reset) ? "in_reset" : "disabled"); atomic_read 313 drivers/misc/habanalabs/memory.c if (atomic_read(&phys_pg_pack->mapping_cnt) > 0) { atomic_read 1030 drivers/misc/habanalabs/memory.c if (atomic_read(&phys_pg_pack->mapping_cnt) == 0) { atomic_read 1163 drivers/misc/habanalabs/memory.c atomic_read(&hdev->in_reset) ? "in_reset" : "disabled"); atomic_read 250 drivers/misc/habanalabs/sysfs.c if (atomic_read(&hdev->in_reset)) atomic_read 47 drivers/misc/ibmasm/command.c dbg("command count: %d\n", atomic_read(&command_count)); atomic_read 58 drivers/misc/ibmasm/command.c dbg("command count: %d\n", atomic_read(&command_count)); atomic_read 493 drivers/misc/lis3lv02d/lis3lv02d.c if (atomic_read(&lis3->wake_thread)) atomic_read 647 drivers/misc/lis3lv02d/lis3lv02d.c if (atomic_read(&lis3->count)) atomic_read 170 drivers/misc/mic/scif/scif_dma.c if ((atomic_read(&ep->rma_info.tcw_total_pages) atomic_read 176 drivers/misc/mic/scif/scif_dma.c atomic_read(&ep->rma_info.tcw_total_pages), atomic_read 194 drivers/misc/mic/scif/scif_nm.c (atomic_read(&scifdev->disconn_rescnt) atomic_read 209 drivers/misc/mic/scif/scif_nm.c (atomic_read(&scifdev->disconn_rescnt) == 1), atomic_read 65 drivers/misc/mic/scif/scif_rma.c !atomic_read(&ep->rma_info.tw_refcount) && atomic_read 66 drivers/misc/mic/scif/scif_rma.c !atomic_read(&ep->rma_info.tcw_refcount) && atomic_read 67 drivers/misc/mic/scif/scif_rma.c !atomic_read(&ep->rma_info.fence_refcount)) atomic_read 260 drivers/misc/phantom.c pr_debug("phantom_poll: %d\n", atomic_read(&dev->counter)); atomic_read 265 drivers/misc/phantom.c else if (atomic_read(&dev->counter)) atomic_read 268 drivers/misc/phantom.c pr_debug("phantom_poll end: %x/%d\n", mask, atomic_read(&dev->counter)); atomic_read 410 drivers/misc/sgi-gru/grufault.c if (atomic_read(>s->ts_gms->ms_range_active)) atomic_read 623 drivers/misc/sgi-gru/grufault.c atomic_read(&gms->ms_range_active) == 0); atomic_read 218 drivers/misc/sgi-gru/grutlbpurge.c range->start, range->end, atomic_read(&gms->ms_range_active)); atomic_read 94 drivers/misc/sgi-xp/xpc_channel.c if (atomic_read(&ch->kthreads_assigned) > 0 || atomic_read 95 drivers/misc/sgi-xp/xpc_channel.c atomic_read(&ch->references) > 0) { atomic_read 123 drivers/misc/sgi-xp/xpc_channel.c if (atomic_read(&ch->n_to_notify) > 0) { atomic_read 136 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(atomic_read(&ch->n_to_notify) != 0); atomic_read 493 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(atomic_read(&ch->kthreads_assigned) != 0); atomic_read 494 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(atomic_read(&ch->kthreads_idle) != 0); atomic_read 495 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(atomic_read(&ch->kthreads_active) != 0); atomic_read 784 drivers/misc/sgi-xp/xpc_channel.c if (atomic_read(&ch->kthreads_idle) > 0) { atomic_read 794 drivers/misc/sgi-xp/xpc_channel.c if (atomic_read(&ch->n_on_msg_allocate_wq) > 0) atomic_read 342 drivers/misc/sgi-xp/xpc_main.c atomic_read(&part->nchannels_active) > 0 || atomic_read 362 drivers/misc/sgi-xp/xpc_main.c (atomic_read(&part->channel_mgr_requests) > 0 || atomic_read 365 drivers/misc/sgi-xp/xpc_main.c atomic_read(&part->nchannels_active) == 0 && atomic_read 494 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); atomic_read 495 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(atomic_read(&part->nchannels_active) != 0); atomic_read 505 drivers/misc/sgi-xp/xpc_main.c wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); atomic_read 612 drivers/misc/sgi-xp/xpc_main.c int idle = atomic_read(&ch->kthreads_idle); atomic_read 613 drivers/misc/sgi-xp/xpc_main.c int assigned = atomic_read(&ch->kthreads_assigned); atomic_read 829 drivers/misc/sgi-xp/xpc_main.c if (atomic_read(&ch->kthreads_assigned) < atomic_read 1364 drivers/misc/sgi-xp/xpc_uv.c if (atomic_read(&ch->n_on_msg_allocate_wq) > 0) atomic_read 1470 drivers/misc/sgi-xp/xpc_uv.c if (atomic_read(&ch->kthreads_idle) > 0) atomic_read 1634 drivers/misc/sgi-xp/xpc_uv.c if (atomic_read(&ch->n_to_notify) == 0) atomic_read 362 drivers/misc/vmw_vmci/vmci_doorbell.c atomic_read(&dbell->active) == 1) { atomic_read 75 drivers/misc/vmw_vmci/vmci_guest.c return atomic_read(&vmci_num_guest_devices) != 0; atomic_read 108 drivers/misc/vmw_vmci/vmci_host.c atomic_read(&vmci_host_active_users) > 0); atomic_read 811 drivers/mmc/core/core.c stop = abort ? atomic_read(abort) : 0; atomic_read 1061 drivers/mmc/core/sdio.c if (atomic_read(&card->sdio_funcs_probed) > 1) { atomic_read 147 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 1) { atomic_read 149 drivers/mtd/maps/vmu-flash.c atomic_read(&mdev->busy) == 0, HZ); atomic_read 150 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 1) { atomic_read 173 drivers/mtd/maps/vmu-flash.c (atomic_read(&mdev->busy) == 0 || atomic_read 174 drivers/mtd/maps/vmu-flash.c atomic_read(&mdev->busy) == 2), HZ * 3); atomic_read 180 drivers/mtd/maps/vmu-flash.c if (error || atomic_read(&mdev->busy) == 2) { atomic_read 181 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 2) atomic_read 252 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 1) { atomic_read 254 drivers/mtd/maps/vmu-flash.c atomic_read(&mdev->busy) == 0, HZ); atomic_read 255 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 1) { atomic_read 268 drivers/mtd/maps/vmu-flash.c atomic_read(&mdev->busy) == 0, HZ/10); atomic_read 274 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 2) { atomic_read 655 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 1) { atomic_read 657 drivers/mtd/maps/vmu-flash.c atomic_read(&mdev->busy) == 0, HZ); atomic_read 658 drivers/mtd/maps/vmu-flash.c if (atomic_read(&mdev->busy) == 1) { atomic_read 4966 drivers/net/bonding/bond_main.c WARN_ON(atomic_read(&netpoll_block_tx)); atomic_read 475 drivers/net/can/c_can/c_can.c idx = fls(atomic_read(&priv->tx_active)); atomic_read 829 drivers/net/can/usb/ems_usb.c if (atomic_read(&dev->active_tx_urbs) >= MAX_TX_URBS || atomic_read 790 drivers/net/can/usb/esd_usb2.c if (atomic_read(&priv->active_tx_jobs) >= MAX_TX_URBS) atomic_read 545 drivers/net/can/usb/gs_usb.c if (atomic_read(&dev->active_tx_urbs) >= GS_MAX_TX_URBS) atomic_read 198 drivers/net/can/usb/mcba_usb.c if (!atomic_read(&priv->free_ctx_cnt)) atomic_read 357 drivers/net/can/usb/peak_usb/pcan_usb_core.c if (atomic_read(&dev->active_tx_urbs) >= PCAN_USB_MAX_TX_URBS) atomic_read 674 drivers/net/can/usb/usb_8dev.c else if (atomic_read(&priv->active_tx_urbs) >= MAX_TX_URBS) atomic_read 1576 drivers/net/ethernet/alteon/acenic.c cur_size = atomic_read(&ap->cur_rx_bufs); atomic_read 1586 drivers/net/ethernet/alteon/acenic.c cur_size = atomic_read(&ap->cur_mini_bufs); atomic_read 1597 drivers/net/ethernet/alteon/acenic.c cur_size = atomic_read(&ap->cur_jumbo_bufs); atomic_read 2173 drivers/net/ethernet/alteon/acenic.c cur_size = atomic_read(&ap->cur_rx_bufs); atomic_read 2187 drivers/net/ethernet/alteon/acenic.c cur_size = atomic_read(&ap->cur_mini_bufs); atomic_read 2204 drivers/net/ethernet/alteon/acenic.c cur_size = atomic_read(&ap->cur_jumbo_bufs); atomic_read 243 drivers/net/ethernet/amazon/ena/ena_com.c cnt = (u16)atomic_read(&admin_queue->outstanding_cmds); atomic_read 1459 drivers/net/ethernet/amazon/ena/ena_com.c while (atomic_read(&admin_queue->outstanding_cmds) != 0) { atomic_read 19 drivers/net/ethernet/aquantia/atlantic/aq_utils.h flags_old = atomic_read(flags); atomic_read 29 drivers/net/ethernet/aquantia/atlantic/aq_utils.h flags_old = atomic_read(flags); atomic_read 36 drivers/net/ethernet/aquantia/atlantic/aq_utils.h return atomic_read(flags) & mask; atomic_read 553 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c pmbox->stats.dpc = atomic_read(&self->dpc); atomic_read 1542 drivers/net/ethernet/atheros/atl1c/atl1c_main.c u16 next_to_clean = atomic_read(&tpd_ring->next_to_clean); atomic_read 1914 drivers/net/ethernet/atheros/atl1c/atl1c_main.c next_to_clean = atomic_read(&tpd_ring->next_to_clean); atomic_read 1234 drivers/net/ethernet/atheros/atl1e/atl1e_main.c u16 next_to_clean = atomic_read(&tx_ring->next_to_clean); atomic_read 1558 drivers/net/ethernet/atheros/atl1e/atl1e_main.c next_to_clean = atomic_read(&tx_ring->next_to_clean); atomic_read 1499 drivers/net/ethernet/atheros/atlx/atl1.c value = ((atomic_read(&adapter->tpd_ring.next_to_use) atomic_read 1501 drivers/net/ethernet/atheros/atlx/atl1.c ((atomic_read(&adapter->rrd_ring.next_to_clean) atomic_read 1503 drivers/net/ethernet/atheros/atlx/atl1.c ((atomic_read(&adapter->rfd_ring.next_to_use) atomic_read 1745 drivers/net/ethernet/atheros/atlx/atl1.c tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); atomic_read 1746 drivers/net/ethernet/atheros/atlx/atl1.c rfd_next_to_use = atomic_read(&adapter->rfd_ring.next_to_use); atomic_read 1747 drivers/net/ethernet/atheros/atlx/atl1.c rrd_next_to_clean = atomic_read(&adapter->rrd_ring.next_to_clean); atomic_read 1844 drivers/net/ethernet/atheros/atlx/atl1.c next_next = rfd_next_to_use = atomic_read(&rfd_ring->next_to_use); atomic_read 1915 drivers/net/ethernet/atheros/atlx/atl1.c rrd_next_to_clean = atomic_read(&rrd_ring->next_to_clean); atomic_read 2035 drivers/net/ethernet/atheros/atlx/atl1.c tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); atomic_read 2037 drivers/net/ethernet/atheros/atlx/atl1.c atomic_read(&adapter->rfd_ring.next_to_use); atomic_read 2039 drivers/net/ethernet/atheros/atlx/atl1.c atomic_read(&adapter->rrd_ring.next_to_clean); atomic_read 2061 drivers/net/ethernet/atheros/atlx/atl1.c sw_tpd_next_to_clean = atomic_read(&tpd_ring->next_to_clean); atomic_read 2093 drivers/net/ethernet/atheros/atlx/atl1.c u16 next_to_clean = atomic_read(&tpd_ring->next_to_clean); atomic_read 2094 drivers/net/ethernet/atheros/atlx/atl1.c u16 next_to_use = atomic_read(&tpd_ring->next_to_use); atomic_read 2202 drivers/net/ethernet/atheros/atlx/atl1.c next_to_use = atomic_read(&tpd_ring->next_to_use); atomic_read 2295 drivers/net/ethernet/atheros/atlx/atl1.c u16 next_to_use = (u16) atomic_read(&tpd_ring->next_to_use); atomic_read 2394 drivers/net/ethernet/atheros/atlx/atl1.c (u16) atomic_read(&tpd_ring->next_to_use)); atomic_read 478 drivers/net/ethernet/atheros/atlx/atl2.c txs_write_ptr = (u32) atomic_read(&adapter->txs_write_ptr); atomic_read 490 drivers/net/ethernet/atheros/atlx/atl2.c txd_read_ptr = (u32) atomic_read(&adapter->txd_read_ptr); atomic_read 808 drivers/net/ethernet/atheros/atlx/atl2.c u32 txs_write_ptr = (u32) atomic_read(&adapter->txs_write_ptr); atomic_read 818 drivers/net/ethernet/atheros/atlx/atl2.c u32 txd_read_ptr = (u32)atomic_read(&adapter->txd_read_ptr); atomic_read 399 drivers/net/ethernet/aurora/nb8800.c if (atomic_read(&priv->tx_free) <= NB8800_DESC_LOW) { atomic_read 3327 drivers/net/ethernet/broadcom/bnx2.c if (unlikely(atomic_read(&bp->intr_sem) != 0)) atomic_read 3344 drivers/net/ethernet/broadcom/bnx2.c if (unlikely(atomic_read(&bp->intr_sem) != 0)) atomic_read 3380 drivers/net/ethernet/broadcom/bnx2.c if (unlikely(atomic_read(&bp->intr_sem) != 0)) atomic_read 6192 drivers/net/ethernet/broadcom/bnx2.c if (atomic_read(&bp->intr_sem) != 0) atomic_read 6561 drivers/net/ethernet/broadcom/bnx2.c atomic_read(&bp->intr_sem), val1); atomic_read 3182 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (atomic_read(&bp->pdev->enable_cnt) != 1) atomic_read 1876 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c DP(BNX2X_MSG_SP, "bp->cq_spq_left %x\n", atomic_read(&bp->cq_spq_left)); atomic_read 3876 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!atomic_read(&bp->eq_spq_left)) { atomic_read 3882 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c } else if (!atomic_read(&bp->cq_spq_left)) { atomic_read 3930 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c atomic_read(&bp->cq_spq_left), atomic_read(&bp->eq_spq_left)); atomic_read 5470 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c hw_cons, sw_cons, atomic_read(&bp->eq_spq_left)); atomic_read 5670 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (atomic_read(&bp->interrupt_occurred)) { atomic_read 13221 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (atomic_read(&pdev->enable_cnt) == 1) { atomic_read 13392 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (atomic_read(&pdev->enable_cnt) == 1) atomic_read 14146 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (atomic_read(&pdev->enable_cnt) == 1) atomic_read 14238 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (atomic_read(&pdev->enable_cnt) == 1) atomic_read 14596 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!atomic_read(&bp->cq_spq_left)) atomic_read 14601 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!atomic_read(&bp->eq_spq_left)) atomic_read 4166 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c c = atomic_read(v); atomic_read 4194 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c c = atomic_read(v); atomic_read 4238 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c cur_credit = atomic_read(&o->credit); atomic_read 2144 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (unlikely(atomic_read(&bp->intr_sem) != 0)) atomic_read 9998 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (atomic_read(&bp->intr_sem) != 0) atomic_read 1543 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c num_flows = atomic_read(&tc_info->flow_table.nelems); atomic_read 90 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c while (atomic_read(&ulp->ref_count) != 0 && i < 10) { atomic_read 548 drivers/net/ethernet/broadcom/cnic.c while ((atomic_read(&ulp_ops->ref_count) != 0) && (i < 20)) { atomic_read 553 drivers/net/ethernet/broadcom/cnic.c if (atomic_read(&ulp_ops->ref_count) != 0) atomic_read 3599 drivers/net/ethernet/broadcom/cnic.c if (atomic_read(&csk1->ref_count)) atomic_read 3654 drivers/net/ethernet/broadcom/cnic.c while (atomic_read(&csk->ref_count) != 1) atomic_read 4230 drivers/net/ethernet/broadcom/cnic.c if (atomic_read(&cp->iscsi_conn) != 0) atomic_read 4232 drivers/net/ethernet/broadcom/cnic.c atomic_read(&cp->iscsi_conn)); atomic_read 5439 drivers/net/ethernet/broadcom/cnic.c while ((atomic_read(&dev->ref_count) != 0) && i < 10) { atomic_read 5443 drivers/net/ethernet/broadcom/cnic.c if (atomic_read(&dev->ref_count) != 0) atomic_read 562 drivers/net/ethernet/brocade/bna/bnad_debugfs.c if (atomic_read(&bna_debugfs_port_count) == 0) { atomic_read 1432 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c if (atomic_read(oct->adapter_refcount) > 1) atomic_read 1501 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c } while ((atomic_read(&ctx.status) == 0) && (count++ < timeout)); atomic_read 1503 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c ret = atomic_read(&ctx.status); atomic_read 442 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c } while ((!atomic_read(&status)) && (count++ < 100000)); atomic_read 444 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c ret = atomic_read(&status); atomic_read 765 drivers/net/ethernet/cavium/liquidio/lio_core.c if (atomic_read(&iq->instr_pending)) atomic_read 996 drivers/net/ethernet/cavium/liquidio/lio_core.c if (!(atomic_read(&oct->status) == OCT_DEV_IN_RESET)) atomic_read 1273 drivers/net/ethernet/cavium/liquidio/lio_core.c atomic_read(&oct->droq[idx]->pkts_pending); atomic_read 269 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(&iq->instr_pending)) { atomic_read 274 drivers/net/ethernet/cavium/liquidio/lio_main.c atomic_read(&iq->instr_pending); atomic_read 911 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(oct_dev->adapter_refcount) == 1) { atomic_read 989 drivers/net/ethernet/cavium/liquidio/lio_main.c switch (atomic_read(&oct->status)) { atomic_read 1037 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(&iq->instr_pending)) { atomic_read 1042 drivers/net/ethernet/cavium/liquidio/lio_main.c atomic_read(&iq->instr_pending); atomic_read 1154 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(oct->adapter_fw_state) == FW_IS_PRELOADED) atomic_read 1261 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING) atomic_read 1280 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(&lio->ifstate) & LIO_IFSTATE_REGISTERED) atomic_read 2348 drivers/net/ethernet/cavium/liquidio/lio_main.c if (!(atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING) || atomic_read 3474 drivers/net/ethernet/cavium/liquidio/lio_main.c } else if (atomic_read(octeon_dev->adapter_fw_state) == atomic_read 3970 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(&oct->status) == OCT_DEV_RUNNING) atomic_read 3978 drivers/net/ethernet/cavium/liquidio/lio_main.c if (atomic_read(&oct->status) != OCT_DEV_CORE_OK) { atomic_read 129 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c if (atomic_read(&iq->instr_pending)) { atomic_read 134 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c atomic_read(&iq->instr_pending); atomic_read 452 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c switch (atomic_read(&oct->status)) { atomic_read 493 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c if (atomic_read(&iq->instr_pending)) { atomic_read 498 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c atomic_read(&iq->instr_pending); atomic_read 674 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c if (atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING) atomic_read 692 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c if (atomic_read(&lio->ifstate) & LIO_IFSTATE_REGISTERED) atomic_read 1433 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c if (!(atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING) || atomic_read 132 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c atomic_set(&vf_rep->ifstate, (atomic_read(&vf_rep->ifstate) | atomic_read 165 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c atomic_set(&vf_rep->ifstate, (atomic_read(&vf_rep->ifstate) & atomic_read 317 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c if (!(atomic_read(&vf_rep->ifstate) & LIO_IFSTATE_RUNNING) || atomic_read 382 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c if (!(atomic_read(&vf_rep->ifstate) & LIO_IFSTATE_RUNNING) || atomic_read 631 drivers/net/ethernet/cavium/liquidio/octeon_device.c s32 istate = (s32)atomic_read(state_ptr); atomic_read 801 drivers/net/ethernet/cavium/liquidio/octeon_device.c refcount = atomic_read(oct->adapter_refcount); atomic_read 818 drivers/net/ethernet/cavium/liquidio/octeon_device.c refcount = atomic_read(oct->adapter_refcount); atomic_read 1204 drivers/net/ethernet/cavium/liquidio/octeon_device.c if (atomic_read(&oct->status) >= OCT_DEV_RUNNING) { atomic_read 1206 drivers/net/ethernet/cavium/liquidio/octeon_device.c atomic_read(&oct->status)); atomic_read 1441 drivers/net/ethernet/cavium/liquidio/octeon_device.c pkts_pend = (u32)atomic_read(&droq->pkts_pending); atomic_read 736 drivers/net/ethernet/cavium/liquidio/octeon_droq.c if (!atomic_read(&droq->pkts_pending)) atomic_read 752 drivers/net/ethernet/cavium/liquidio/octeon_droq.c pkt_count = atomic_read(&droq->pkts_pending); atomic_read 774 drivers/net/ethernet/cavium/liquidio/octeon_droq.c if (atomic_read(&droq->pkts_pending)) atomic_read 800 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (u32)(atomic_read(&droq->pkts_pending))); atomic_read 506 drivers/net/ethernet/cavium/liquidio/octeon_network.h return atomic_read(&lio->ifstate) & state_flag; atomic_read 516 drivers/net/ethernet/cavium/liquidio/octeon_network.h atomic_set(&lio->ifstate, (atomic_read(&lio->ifstate) | state_flag)); atomic_read 526 drivers/net/ethernet/cavium/liquidio/octeon_network.h atomic_set(&lio->ifstate, (atomic_read(&lio->ifstate) & ~(state_flag))); atomic_read 540 drivers/net/ethernet/cavium/liquidio/octeon_network.h pcount = atomic_read( atomic_read 114 drivers/net/ethernet/cavium/liquidio/octeon_nic.h return ((u32)atomic_read(&oct->instr_queue[q_no]->instr_pending) atomic_read 259 drivers/net/ethernet/cavium/liquidio/request_manager.c atomic_read(&oct->instr_queue[i]->instr_pending); atomic_read 278 drivers/net/ethernet/cavium/liquidio/request_manager.c if (atomic_read(&oct->status) == OCT_DEV_RUNNING) { atomic_read 320 drivers/net/ethernet/cavium/liquidio/request_manager.c if (atomic_read(&iq->instr_pending) >= (s32)(iq->max_count - 1)) { atomic_read 326 drivers/net/ethernet/cavium/liquidio/request_manager.c if (atomic_read(&iq->instr_pending) >= (s32)(iq->max_count - 2)) atomic_read 440 drivers/net/ethernet/cavium/liquidio/request_manager.c if (atomic_read(&oct->response_list atomic_read 514 drivers/net/ethernet/cavium/liquidio/request_manager.c if (!atomic_read(&iq->instr_pending)) atomic_read 782 drivers/net/ethernet/cavium/liquidio/request_manager.c if (!atomic_read(&done_sc_list->pending_req_count)) atomic_read 231 drivers/net/ethernet/cavium/liquidio/response_manager.c if (atomic_read(&oct->response_list atomic_read 937 drivers/net/ethernet/cavium/thunder/nicvf_main.c (atomic_read(&sq->free_cnt) >= MIN_SQ_DESC_PER_PKT_XMIT)) { atomic_read 1300 drivers/net/ethernet/cavium/thunder/nicvf_main.c if (atomic_read(&sq->free_cnt) > MIN_SQ_DESC_PER_PKT_XMIT) { atomic_read 1558 drivers/net/ethernet/cavium/thunder/nicvf_queues.c if (subdesc_cnt > atomic_read(&sq->free_cnt)) atomic_read 658 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c t->atids_in_use + atomic_read(&t->tids_in_use) + MC5_MIN_TIDS <= atomic_read 223 drivers/net/ethernet/chelsio/cxgb3/l2t.c if (!atomic_read(&d->nfree)) atomic_read 228 drivers/net/ethernet/chelsio/cxgb3/l2t.c if (atomic_read(&e->refcnt) == 0) atomic_read 231 drivers/net/ethernet/chelsio/cxgb3/l2t.c for (e = &d->l2tab[1]; atomic_read(&e->refcnt); ++e) ; atomic_read 267 drivers/net/ethernet/chelsio/cxgb3/l2t.c if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */ atomic_read 338 drivers/net/ethernet/chelsio/cxgb3/l2t.c if (atomic_read(&e->refcnt) == 1) atomic_read 421 drivers/net/ethernet/chelsio/cxgb3/l2t.c if (atomic_read(&e->refcnt)) { atomic_read 271 drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c seq_printf(seq, "Free clip entries : %d\n", atomic_read(&ctbl->nfree)); atomic_read 3135 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->conns_in_use)); atomic_read 3147 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->tids_in_use), atomic_read 3148 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->hash_tids_in_use)); atomic_read 3156 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->tids_in_use), atomic_read 3157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->hash_tids_in_use)); atomic_read 3163 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->hash_tids_in_use)); atomic_read 3167 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->conns_in_use)); atomic_read 3172 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&t->tids_in_use)); atomic_read 3339 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.cipher_rqst)); atomic_read 3341 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.digest_rqst)); atomic_read 3343 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.aead_rqst)); atomic_read 3345 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.complete)); atomic_read 3347 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.error)); atomic_read 3349 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.fallback)); atomic_read 3351 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c atomic_read(&adap->chcr_stats.ipsec_cnt)); atomic_read 268 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (!atomic_read(&d->nfree)) atomic_read 273 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (atomic_read(&e->refcnt) == 0) atomic_read 276 drivers/net/ethernet/chelsio/cxgb4/l2t.c for (e = d->l2tab; atomic_read(&e->refcnt); ++e) atomic_read 305 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (atomic_read(&e->refcnt) == 0) { atomic_read 356 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */ atomic_read 376 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */ atomic_read 448 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (atomic_read(&e->refcnt) == 1) atomic_read 548 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (atomic_read(&e->refcnt)) atomic_read 596 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (!atomic_read(&e->refcnt)) { atomic_read 725 drivers/net/ethernet/chelsio/cxgb4/l2t.c l2e_state(e), atomic_read(&e->refcnt), atomic_read 586 drivers/net/ethernet/emulex/benet/be_cmds.c if (atomic_read(&mcc_obj->q.used) == 0) atomic_read 847 drivers/net/ethernet/emulex/benet/be_cmds.c if (atomic_read(&mccq->used) >= mccq->len) atomic_read 819 drivers/net/ethernet/emulex/benet/be_main.c return atomic_read(&txo->q.used) + BE_MAX_TX_FRAG_COUNT >= txo->q.len; atomic_read 824 drivers/net/ethernet/emulex/benet/be_main.c return atomic_read(&txo->q.used) < txo->q.len / 2; atomic_read 829 drivers/net/ethernet/emulex/benet/be_main.c return atomic_read(&txo->q.used) > txo->pend_wrb_cnt; atomic_read 1435 drivers/net/ethernet/emulex/benet/be_main.c atomic_read(&txo->q.used), txo->q.id); atomic_read 1450 drivers/net/ethernet/emulex/benet/be_main.c atomic_read(&txo->cq.used)); atomic_read 2662 drivers/net/ethernet/emulex/benet/be_main.c } else if (atomic_read(&rxq->used) == 0) { atomic_read 2817 drivers/net/ethernet/emulex/benet/be_main.c while (atomic_read(&rxq->used) > 0) { atomic_read 2822 drivers/net/ethernet/emulex/benet/be_main.c BUG_ON(atomic_read(&rxq->used)); atomic_read 2911 drivers/net/ethernet/emulex/benet/be_main.c if (atomic_read(&txq->used)) { atomic_read 2913 drivers/net/ethernet/emulex/benet/be_main.c i, atomic_read(&txq->used)); atomic_read 2916 drivers/net/ethernet/emulex/benet/be_main.c index_adv(&end_idx, atomic_read(&txq->used) - 1, atomic_read 2923 drivers/net/ethernet/emulex/benet/be_main.c BUG_ON(atomic_read(&txq->used)); atomic_read 3255 drivers/net/ethernet/emulex/benet/be_main.c if (atomic_read(&rxo->q.used) < RX_FRAGS_REFILL_WM && atomic_read 3582 drivers/net/ethernet/emulex/benet/be_main.c if (atomic_read(&q->used) == 0) atomic_read 47 drivers/net/ethernet/google/gve/gve_tx.c WARN(atomic_read(&fifo->available) != fifo->size, atomic_read 61 drivers/net/ethernet/google/gve/gve_tx.c return (atomic_read(&fifo->available) <= bytes) ? false : true; atomic_read 218 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c if (atomic_read(&hdev->arq.count) >= atomic_read 724 drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c HINIC_CMDQ_CTXT_BLOCK_INFO_SET(atomic_read(&wq->cons_idx), CI); atomic_read 101 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c ci_start = atomic_read(&wq->cons_idx); atomic_read 102 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c pi_start = atomic_read(&wq->prod_idx); atomic_read 158 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c ci_start = atomic_read(&wq->cons_idx); atomic_read 159 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c pi_start = atomic_read(&wq->prod_idx); atomic_read 455 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c return atomic_read(&wq->delta) - 1; atomic_read 468 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c return atomic_read(&wq->delta) - 1; atomic_read 748 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c *prod_idx = MASKED_WQE_IDX(wq, atomic_read(&wq->prod_idx)); atomic_read 828 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c if ((atomic_read(&wq->delta) + num_wqebbs) > wq->q_depth) atomic_read 831 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c curr_cons_idx = atomic_read(&wq->cons_idx); atomic_read 258 drivers/net/ethernet/ibm/ehea/ehea_ethtool.c data[i++] = atomic_read(&port->port_res[k].swqe_avail); atomic_read 858 drivers/net/ethernet/ibm/ehea/ehea_main.c (atomic_read(&pr->swqe_avail) >= pr->swqe_refill_th))) { atomic_read 861 drivers/net/ethernet/ibm/ehea/ehea_main.c (atomic_read(&pr->swqe_avail) >= pr->swqe_refill_th)) atomic_read 2077 drivers/net/ethernet/ibm/ehea/ehea_main.c if (unlikely(atomic_read(&pr->swqe_avail) <= 1)) { atomic_read 2495 drivers/net/ethernet/ibm/ehea/ehea_main.c atomic_read(&pr->swqe_avail) >= swqe_max, atomic_read 3351 drivers/net/ethernet/ibm/ehea/ehea_main.c if (atomic_read(&ehea_memory_hooks_registered) == 0) atomic_read 209 drivers/net/ethernet/ibm/ibmveth.c u32 count = pool->size - atomic_read(&pool->available); atomic_read 318 drivers/net/ethernet/ibm/ibmveth.c (atomic_read(&pool->available) < pool->threshold)) atomic_read 245 drivers/net/ethernet/ibm/ibmvnic.c int count = pool->size - atomic_read(&pool->available); atomic_read 4122 drivers/net/ethernet/ibm/ibmvnic.c if (atomic_read(&adapter->running_cap_crqs) == 0) { atomic_read 4234 drivers/net/ethernet/ibm/ibmvnic.c atomic_read(&adapter->running_cap_crqs)); atomic_read 4398 drivers/net/ethernet/ibm/ibmvnic.c if (atomic_read(&adapter->running_cap_crqs) == 0) { atomic_read 4665 drivers/net/ethernet/ibm/ibmvnic.c if (atomic_read(&adapter->running_cap_crqs) != 0) atomic_read 2356 drivers/net/ethernet/intel/e1000/e1000_main.c if (atomic_read(&adapter->tx_fifo_stall)) { atomic_read 3048 drivers/net/ethernet/intel/e1000/e1000_main.c if (atomic_read(&adapter->tx_fifo_stall)) atomic_read 1063 drivers/net/ethernet/jme.c if (unlikely(atomic_read(&jme->link_changing) != 1)) atomic_read 1069 drivers/net/ethernet/jme.c i = atomic_read(&rxring->next_to_clean); atomic_read 1201 drivers/net/ethernet/jme.c (atomic_read(&jme->link_changing) != 1) atomic_read 1278 drivers/net/ethernet/jme.c while (atomic_read(&jme->link_changing) != 1) atomic_read 1366 drivers/net/ethernet/jme.c while (atomic_read(&jme->rx_empty) > 0) { atomic_read 1387 drivers/net/ethernet/jme.c if (unlikely(atomic_read(&jme->link_changing) != 1)) atomic_read 1397 drivers/net/ethernet/jme.c while (atomic_read(&jme->rx_empty) > 0) { atomic_read 1412 drivers/net/ethernet/jme.c atomic_read(&txring->nr_free) >= (jme->tx_wake_threshold))) { atomic_read 1433 drivers/net/ethernet/jme.c if (unlikely(atomic_read(&jme->link_changing) != 1)) atomic_read 1439 drivers/net/ethernet/jme.c max = jme->tx_ring_size - atomic_read(&txring->nr_free); atomic_read 1442 drivers/net/ethernet/jme.c for (i = atomic_read(&txring->next_to_clean) ; cnt < max ; ) { atomic_read 1955 drivers/net/ethernet/jme.c if (unlikely(atomic_read(&txring->nr_free) < nr_alloc)) atomic_read 2192 drivers/net/ethernet/jme.c int idx = atomic_read(&txring->next_to_clean); atomic_read 2197 drivers/net/ethernet/jme.c if (unlikely(atomic_read(&txring->nr_free) < (MAX_SKB_FRAGS+2))) { atomic_read 2201 drivers/net/ethernet/jme.c if (atomic_read(&txring->nr_free) atomic_read 1145 drivers/net/ethernet/mediatek/mtk_eth_soc.c if (unlikely(atomic_read(&ring->free_count) <= tx_num)) { atomic_read 1171 drivers/net/ethernet/mediatek/mtk_eth_soc.c if (unlikely(atomic_read(&ring->free_count) <= ring->thresh)) atomic_read 1450 drivers/net/ethernet/mediatek/mtk_eth_soc.c (atomic_read(&ring->free_count) > ring->thresh)) atomic_read 2704 drivers/net/ethernet/mellanox/mlx4/fw.c int num_tasks = atomic_read(&priv->opreq_count); atomic_read 2050 drivers/net/ethernet/mellanox/mlx4/main.c if (atomic_read(&pf_loading)) { atomic_read 1338 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (res->com.state == RES_QP_BUSY || atomic_read(&res->ref_count) || atomic_read 1341 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c res->com.state, atomic_read(&res->ref_count)); atomic_read 1353 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c atomic_read(&res->ref_count)) { atomic_read 1357 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c atomic_read(&res->ref_count)); atomic_read 1688 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c else if (atomic_read(&r->ref_count)) atomic_read 1728 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c else if (atomic_read(&r->ref_count)) atomic_read 4818 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (cq->com.owner == slave && !atomic_read(&cq->ref_count)) { atomic_read 1774 drivers/net/ethernet/mellanox/mlx5/core/cmd.c wait_event(ctx->wait, atomic_read(&ctx->num_inflight) == 0); atomic_read 4172 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return atomic_read(&tc_ht->nelems); atomic_read 359 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (atomic_read(&rep->rep_data[REP_ETH].state) != REP_LOADED) atomic_read 2571 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (atomic_read(&rep->rep_data[rep_type].state) == REP_LOADED && atomic_read 1624 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c version += (u64)atomic_read(&iter->g->node.version); atomic_read 1709 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c if (atomic_read(&ft->node.version) != ft_version) { atomic_read 1779 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c version = atomic_read(&ft->node.version); atomic_read 1810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c version != atomic_read(&ft->node.version)) atomic_read 1596 drivers/net/ethernet/natsemi/ns83820.c dev->tx_done_idx, dev->tx_free_idx, atomic_read(&dev->nr_tx_skbs) atomic_read 1605 drivers/net/ethernet/natsemi/ns83820.c atomic_read(&dev->nr_tx_skbs)); atomic_read 300 drivers/net/ethernet/netronome/nfp/flower/main.c atomic_read(replies) >= tot_repl, atomic_read 504 drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c *data++ = atomic_read(&nn->ktls_no_space); atomic_read 408 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c if (!atomic_read(&bar->refcnt)) atomic_read 805 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c if (WARN_ON(!atomic_read(&priv->refcnt))) atomic_read 405 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c if (atomic_read(&area->refcount)) atomic_read 136 drivers/net/ethernet/qlogic/qed/qed_main.c if (atomic_read(&pdev->enable_cnt) == 1) atomic_read 172 drivers/net/ethernet/qlogic/qed/qed_main.c if (atomic_read(&pdev->enable_cnt) == 1) { atomic_read 2471 drivers/net/ethernet/qlogic/qla3xxx.c if (unlikely(atomic_read(&qdev->tx_count) < 2)) atomic_read 836 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c if (atomic_read(&cmd->rsp_status) == atomic_read 2036 drivers/net/ethernet/sfc/ef10.c if (rc != -ENOENT || atomic_read(&efx->active_queues)) atomic_read 3952 drivers/net/ethernet/sfc/ef10.c WARN_ON(atomic_read(&efx->active_queues) < 0); atomic_read 3981 drivers/net/ethernet/sfc/ef10.c atomic_read(&efx->active_queues) == 0, atomic_read 3983 drivers/net/ethernet/sfc/ef10.c pending = atomic_read(&efx->active_queues); atomic_read 3110 drivers/net/ethernet/sfc/efx.c stats[GENERIC_STAT_rx_noskb_drops] = atomic_read(&efx->n_rx_noskb_drops); atomic_read 49 drivers/net/ethernet/sfc/ethtool.c return atomic_read((atomic_t *) field); atomic_read 2719 drivers/net/ethernet/sfc/falcon/efx.c stats[GENERIC_STAT_rx_noskb_drops] = atomic_read(&efx->n_rx_noskb_drops); atomic_read 49 drivers/net/ethernet/sfc/falcon/ethtool.c return atomic_read((atomic_t *) field); atomic_read 434 drivers/net/ethernet/sfc/falcon/farch.c WARN_ON(atomic_read(&tx_queue->flush_outstanding)); atomic_read 609 drivers/net/ethernet/sfc/falcon/farch.c return (atomic_read(&efx->active_queues) == 0 || atomic_read 610 drivers/net/ethernet/sfc/falcon/farch.c (atomic_read(&efx->rxq_flush_outstanding) < EF4_RX_FLUSH_COUNT atomic_read 611 drivers/net/ethernet/sfc/falcon/farch.c && atomic_read(&efx->rxq_flush_pending) > 0)); atomic_read 676 drivers/net/ethernet/sfc/falcon/farch.c while (timeout && atomic_read(&efx->active_queues) > 0) { atomic_read 683 drivers/net/ethernet/sfc/falcon/farch.c if (atomic_read(&efx->rxq_flush_outstanding) >= atomic_read 701 drivers/net/ethernet/sfc/falcon/farch.c if (atomic_read(&efx->active_queues) && atomic_read 704 drivers/net/ethernet/sfc/falcon/farch.c "(rx %d+%d)\n", atomic_read(&efx->active_queues), atomic_read 705 drivers/net/ethernet/sfc/falcon/farch.c atomic_read(&efx->rxq_flush_outstanding), atomic_read 706 drivers/net/ethernet/sfc/falcon/farch.c atomic_read(&efx->rxq_flush_pending)); atomic_read 1147 drivers/net/ethernet/sfc/falcon/farch.c WARN_ON(atomic_read(&efx->active_queues) == 0); atomic_read 362 drivers/net/ethernet/sfc/falcon/selftest.c if (atomic_read(&state->rx_bad) == 0) { atomic_read 463 drivers/net/ethernet/sfc/falcon/selftest.c return atomic_read(&state->rx_good) == state->packet_count; atomic_read 489 drivers/net/ethernet/sfc/falcon/selftest.c rx_good = atomic_read(&state->rx_good); atomic_read 490 drivers/net/ethernet/sfc/falcon/selftest.c rx_bad = atomic_read(&state->rx_bad); atomic_read 422 drivers/net/ethernet/sfc/farch.c WARN_ON(atomic_read(&tx_queue->flush_outstanding)); atomic_read 591 drivers/net/ethernet/sfc/farch.c return (atomic_read(&efx->active_queues) == 0 || atomic_read 592 drivers/net/ethernet/sfc/farch.c (atomic_read(&efx->rxq_flush_outstanding) < EFX_RX_FLUSH_COUNT atomic_read 593 drivers/net/ethernet/sfc/farch.c && atomic_read(&efx->rxq_flush_pending) > 0)); atomic_read 658 drivers/net/ethernet/sfc/farch.c while (timeout && atomic_read(&efx->active_queues) > 0) { atomic_read 675 drivers/net/ethernet/sfc/farch.c if (atomic_read(&efx->rxq_flush_outstanding) >= atomic_read 694 drivers/net/ethernet/sfc/farch.c if (atomic_read(&efx->active_queues) && atomic_read 697 drivers/net/ethernet/sfc/farch.c "(rx %d+%d)\n", atomic_read(&efx->active_queues), atomic_read 698 drivers/net/ethernet/sfc/farch.c atomic_read(&efx->rxq_flush_outstanding), atomic_read 699 drivers/net/ethernet/sfc/farch.c atomic_read(&efx->rxq_flush_pending)); atomic_read 1135 drivers/net/ethernet/sfc/farch.c WARN_ON(atomic_read(&efx->active_queues) == 0); atomic_read 1276 drivers/net/ethernet/sfc/mcdi_port.c if (rc && (rc != -ENOENT || atomic_read(&efx->active_queues))) atomic_read 362 drivers/net/ethernet/sfc/selftest.c if (atomic_read(&state->rx_bad) == 0) { atomic_read 463 drivers/net/ethernet/sfc/selftest.c return atomic_read(&state->rx_good) == state->packet_count; atomic_read 489 drivers/net/ethernet/sfc/selftest.c rx_good = atomic_read(&state->rx_good); atomic_read 490 drivers/net/ethernet/sfc/selftest.c rx_bad = atomic_read(&state->rx_bad); atomic_read 663 drivers/net/ethernet/sfc/siena_sriov.c atomic_read(&vf->rxq_retry_count); atomic_read 864 drivers/net/ethernet/silan/sc92031.c intr_mask = atomic_read(&priv->intr_mask); atomic_read 897 drivers/net/ethernet/silan/sc92031.c intr_mask = atomic_read(&priv->intr_mask); atomic_read 3834 drivers/net/ethernet/sun/cassini.c while (atomic_read(&cp->reset_task_pending_mtu) || atomic_read 3835 drivers/net/ethernet/sun/cassini.c atomic_read(&cp->reset_task_pending_spare) || atomic_read 3836 drivers/net/ethernet/sun/cassini.c atomic_read(&cp->reset_task_pending_all)) atomic_read 3840 drivers/net/ethernet/sun/cassini.c while (atomic_read(&cp->reset_task_pending)) atomic_read 3997 drivers/net/ethernet/sun/cassini.c int pending = atomic_read(&cp->reset_task_pending); atomic_read 3999 drivers/net/ethernet/sun/cassini.c int pending_all = atomic_read(&cp->reset_task_pending_all); atomic_read 4000 drivers/net/ethernet/sun/cassini.c int pending_spare = atomic_read(&cp->reset_task_pending_spare); atomic_read 4001 drivers/net/ethernet/sun/cassini.c int pending_mtu = atomic_read(&cp->reset_task_pending_mtu); atomic_read 4098 drivers/net/ethernet/sun/cassini.c if (atomic_read(&cp->reset_task_pending_all) || atomic_read 4099 drivers/net/ethernet/sun/cassini.c atomic_read(&cp->reset_task_pending_spare) || atomic_read 4100 drivers/net/ethernet/sun/cassini.c atomic_read(&cp->reset_task_pending_mtu)) atomic_read 4103 drivers/net/ethernet/sun/cassini.c if (atomic_read(&cp->reset_task_pending)) atomic_read 542 drivers/net/ethernet/ti/cpmac.c if (unlikely(atomic_read(&priv->reset_pending))) atomic_read 1829 drivers/net/ethernet/toshiba/ps3_gelic_net.c atomic_read(&card->tx_timeout_task_counter) == 0); atomic_read 2475 drivers/net/ethernet/toshiba/spider_net.c atomic_read(&card->tx_timeout_task_counter) == 0); atomic_read 875 drivers/net/hyperv/netvsc.c atomic_read(&nvchan->queue_sends) < 1 && atomic_read 1006 drivers/net/hyperv/rndis_filter.c if (atomic_read(&nvchan->queue_sends) > 0) atomic_read 1167 drivers/net/hyperv/rndis_filter.c atomic_read(&nvdev->open_chn) == nvdev->num_chn); atomic_read 383 drivers/net/plip/plip.c if (!(atomic_read (&nl->kill_timer))) { atomic_read 3271 drivers/net/ppp/ppp_generic.c if (atomic_read(&ppp_unit_count) || atomic_read(&channel_count)) atomic_read 418 drivers/net/rionet.c state = atomic_read(&rdev->state); atomic_read 1016 drivers/net/thunderbolt.c u16 frame_id = atomic_read(&net->frame_id); atomic_read 263 drivers/net/usb/catc.c if (atomic_read(&catc->recq_sz)) { atomic_read 48 drivers/net/usb/cdc_mbim.c dev_dbg(&dev->intf->dev, "%s() pmcount=%d, on=%d\n", __func__, atomic_read(&info->pmcount), on); atomic_read 1347 drivers/net/usb/cdc_ncm.c if (!(hrtimer_active(&ctx->tx_timer) || atomic_read(&ctx->stop))) atomic_read 1358 drivers/net/usb/cdc_ncm.c if (!atomic_read(&ctx->stop)) atomic_read 637 drivers/net/usb/qmi_wwan.c atomic_read(&info->pmcount), on); atomic_read 1590 drivers/net/usb/r8152.c WARN_ON(atomic_read(&tp->rx_count)); atomic_read 2003 drivers/net/usb/r8152.c return atomic_read(&tp->rx_count) > RTL8152_MAX_RX; atomic_read 2035 drivers/net/usb/r8152.c if (!agg_free && atomic_read(&tp->rx_count) < tp->rx_pending) atomic_read 621 drivers/net/wimax/i2400m/driver.c if (atomic_read(&i2400m->bus_reset_retries) atomic_read 201 drivers/net/wimax/i2400m/usb-rx.c do_autopm = atomic_read(&i2400mu->do_autopm); atomic_read 351 drivers/net/wimax/i2400m/usb-rx.c || (pending = atomic_read(&i2400mu->rx_pending_count))) atomic_read 97 drivers/net/wimax/i2400m/usb-tx.c do_autopm = atomic_read(&i2400mu->do_autopm); atomic_read 137 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_pending)); atomic_read 730 drivers/net/wireless/ath/ar5523/ar5523.c if (atomic_read(&ar->tx_nr_total) < AR5523_TX_DATA_RESTART_COUNT) { atomic_read 776 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_total), atomic_read 777 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_pending)); atomic_read 867 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_pending)); atomic_read 911 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_total), atomic_read 912 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_pending)); atomic_read 927 drivers/net/wireless/ath/ar5523/ar5523.c !atomic_read(&ar->tx_nr_pending), AR5523_FLUSH_TIMEOUT)) atomic_read 929 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_total), atomic_read 930 drivers/net/wireless/ath/ar5523/ar5523.c atomic_read(&ar->tx_nr_pending)); atomic_read 3984 drivers/net/wireless/ath/ath10k/htt_rx.c while (atomic_read(&htt->num_mpdus_ready)) { atomic_read 611 drivers/net/wireless/ath/ath6kl/sdio.c return !atomic_read(&ar_sdio->irq_handling); atomic_read 621 drivers/net/wireless/ath/ath6kl/sdio.c if (atomic_read(&ar_sdio->irq_handling)) { atomic_read 83 drivers/net/wireless/ath/ath9k/htc_drv_init.c if (atomic_read(&priv->htc->tgt_ready) > 0) { atomic_read 352 drivers/net/wireless/ath/ath9k/htc_drv_init.c if (atomic_read(&priv->wmi->mwrite_cnt)) atomic_read 506 drivers/net/wireless/ath/ath9k/htc_drv_init.c if (atomic_read(&priv->wmi->m_rmw_cnt)) atomic_read 881 drivers/net/wireless/ath/ath9k/mac.c if (atomic_read(&ah->intr_ref_cnt) != 0) { atomic_read 883 drivers/net/wireless/ath/ath9k/mac.c atomic_read(&ah->intr_ref_cnt)); atomic_read 900 drivers/net/wireless/ath/ath9k/mac.c atomic_read(&ah->intr_ref_cnt)); atomic_read 226 drivers/net/wireless/ath/carl9170/debug.c ar->fw.mem_blocks, atomic_read(&ar->mem_allocs)); atomic_read 229 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->mem_free_blocks), atomic_read 230 drivers/net/wireless/ath/carl9170/debug.c (atomic_read(&ar->mem_free_blocks) * ar->fw.mem_block_size) / 1024, atomic_read 677 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->pending_restarts)); atomic_read 773 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->tx_anch_urbs)); atomic_read 775 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->rx_anch_urbs)); atomic_read 777 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->rx_work_urbs)); atomic_read 779 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->rx_pool_urbs)); atomic_read 782 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->tx_total_queued)); atomic_read 784 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->tx_ampdu_scheduler)); atomic_read 787 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->tx_total_pending)); atomic_read 793 drivers/net/wireless/ath/carl9170/debug.c atomic_read(&ar->tx_ampdu_upload)); atomic_read 245 drivers/net/wireless/ath/carl9170/main.c if (atomic_read(&ar->tx_total_queued)) atomic_read 1733 drivers/net/wireless/ath/carl9170/main.c if (atomic_read(&sta_info->pending_frames)) atomic_read 1747 drivers/net/wireless/ath/carl9170/main.c return !!atomic_read(&ar->tx_total_queued); atomic_read 73 drivers/net/wireless/ath/carl9170/tx.c atomic_read(&ar->mem_free_blocks)); atomic_read 285 drivers/net/wireless/ath/carl9170/tx.c if (atomic_read(&ar->tx_total_queued)) atomic_read 289 drivers/net/wireless/ath/carl9170/tx.c if (!atomic_read(&ar->tx_ampdu_upload)) atomic_read 651 drivers/net/wireless/ath/carl9170/tx.c if (!atomic_read(&ar->tx_total_queued)) atomic_read 1139 drivers/net/wireless/ath/carl9170/tx.c if (atomic_read(&ar->tx_ampdu_upload)) atomic_read 321 drivers/net/wireless/ath/carl9170/usb.c while ((atomic_read(&ar->rx_anch_urbs) < AR9170_NUM_RX_URBS) && atomic_read 439 drivers/net/wireless/ath/carl9170/usb.c if (atomic_read(&ar->rx_anch_urbs) == 0) { atomic_read 150 drivers/net/wireless/ath/wil6210/interrupt.c bool unmask_rx_htrsh = atomic_read(&wil->connected_vifs) > 0; atomic_read 820 drivers/net/wireless/broadcom/b43/b43.h #define b43_status(wldev) atomic_read(&(wldev)->__init_status) atomic_read 62 drivers/net/wireless/broadcom/b43/leds.c turn_on = atomic_read(&led->state) != LED_OFF; atomic_read 663 drivers/net/wireless/broadcom/b43legacy/b43legacy.h #define b43legacy_status(wldev) atomic_read(&(wldev)->__init_status) atomic_read 803 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c WARN_ON(atomic_read(&sdiodev->freezer->freezing)); atomic_read 818 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c atomic_read(expect) == sdiodev->freezer->frozen_count); atomic_read 836 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c return atomic_read(&sdiodev->freezer->freezing); atomic_read 1396 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c return atomic_read(&ifp->pend_8021x_cnt); atomic_read 787 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if ((force) || (atomic_read(&commonring->outstanding_tx) < atomic_read 1383 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c ((qlen) && (atomic_read(&commonring->outstanding_tx) < atomic_read 2511 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (!sdiodev->irq_en && !atomic_read(&bus->ipend)) { atomic_read 2591 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (atomic_read(&bus->ipend) > 0) { atomic_read 2676 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if ((bus->clkstate == CLK_AVAIL) && !atomic_read(&bus->fcstate) && atomic_read 2697 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c } else if (atomic_read(&bus->intstatus) || atomic_read 2698 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c atomic_read(&bus->ipend) > 0 || atomic_read 2699 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c (!atomic_read(&bus->fcstate) && atomic_read 315 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c while (atomic_read(&wl->callbacks) > 0) atomic_read 1440 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c callbacks = atomic_read(&wl->callbacks) - ret_val; atomic_read 1448 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c SPINWAIT((atomic_read(&wl->callbacks) > callbacks), 100 * 1000); atomic_read 851 drivers/net/wireless/intel/iwlegacy/debug.c atomic_read(&il->queue_stop_count[cnt])); atomic_read 1557 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c if (atomic_read(&sta_priv->pending_frames) > 0) atomic_read 4903 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c atomic_read(&mvm->queue_sync_counter) == 0 || atomic_read 540 drivers/net/wireless/intel/iwlwifi/pcie/rx.c int pending = atomic_read(&rba->req_pending); atomic_read 599 drivers/net/wireless/intel/iwlwifi/pcie/rx.c pending = atomic_read(&rba->req_pending); atomic_read 1461 drivers/net/wireless/intel/iwlwifi/pcie/rx.c atomic_read(&trans_pcie->rba.req_pending) * atomic_read 192 drivers/net/wireless/intersil/hostap/hostap_ap.c if (atomic_read(&sta->users) != 0) atomic_read 1007 drivers/net/wireless/intersil/hostap/hostap_ap.c sta->addr, atomic_read(&sta->users), sta->aid, atomic_read 145 drivers/net/wireless/marvell/mwifiex/cmdevt.c atomic_read(&adapter->cmd_pending)); atomic_read 727 drivers/net/wireless/marvell/mwifiex/cmdevt.c command, atomic_read(&adapter->cmd_pending)); atomic_read 1121 drivers/net/wireless/marvell/mwifiex/cmdevt.c if (!adapter->cmd_sent && !atomic_read(&adapter->tx_hw_pending) && atomic_read 1128 drivers/net/wireless/marvell/mwifiex/cmdevt.c atomic_read(&adapter->tx_hw_pending) ? "T" : "", atomic_read 274 drivers/net/wireless/marvell/mwifiex/debugfs.c atomic_read(&phist_data->num_samples)); atomic_read 288 drivers/net/wireless/marvell/mwifiex/debugfs.c value = atomic_read(&phist_data->rx_rate[i]); atomic_read 296 drivers/net/wireless/marvell/mwifiex/debugfs.c value = atomic_read(&phist_data->rx_rate[i]); atomic_read 304 drivers/net/wireless/marvell/mwifiex/debugfs.c value = atomic_read(&phist_data->snr[i]); atomic_read 309 drivers/net/wireless/marvell/mwifiex/debugfs.c value = atomic_read(&phist_data->noise_flr[i]); atomic_read 315 drivers/net/wireless/marvell/mwifiex/debugfs.c value = atomic_read(&phist_data->sig_str[i]); atomic_read 204 drivers/net/wireless/marvell/mwifiex/main.c (atomic_read(&adapter->rx_pending) < LOW_RX_PENDING)) { atomic_read 270 drivers/net/wireless/marvell/mwifiex/main.c if (atomic_read(&adapter->rx_pending) >= HIGH_RX_PENDING && atomic_read 1053 drivers/net/wireless/marvell/mwifiex/main.c if (atomic_read(&card->port[i].tx_data_urb_pending)) { atomic_read 1116 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&cardp->tx_cmd_urb_pending)); atomic_read 1118 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&cardp->port[0].tx_data_urb_pending)); atomic_read 1120 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&cardp->port[1].tx_data_urb_pending)); atomic_read 1122 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&cardp->rx_cmd_urb_pending)); atomic_read 1124 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&cardp->rx_data_urb_pending)); atomic_read 1128 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->tx_pending)); atomic_read 1130 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->rx_pending)); atomic_read 1147 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&priv->wmm_tx_pending[0])); atomic_read 1149 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&priv->wmm_tx_pending[1])); atomic_read 1151 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&priv->wmm_tx_pending[2])); atomic_read 1153 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&priv->wmm_tx_pending[3])); atomic_read 1425 drivers/net/wireless/marvell/mwifiex/main.c if (atomic_read(&adapter->rx_pending) || atomic_read 1426 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->tx_pending) || atomic_read 1427 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->cmd_pending)) { atomic_read 1431 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->rx_pending), atomic_read 1432 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->tx_pending), atomic_read 1433 drivers/net/wireless/marvell/mwifiex/main.c atomic_read(&adapter->cmd_pending)); atomic_read 72 drivers/net/wireless/marvell/mwifiex/tdls.c if (atomic_read(&priv->wmm.highest_queued_prio) < atomic_read 59 drivers/net/wireless/marvell/mwifiex/uap_txrx.c if ((atomic_read(&adapter->pending_bridged_pkts) <= atomic_read 109 drivers/net/wireless/marvell/mwifiex/uap_txrx.c if ((atomic_read(&adapter->pending_bridged_pkts) >= atomic_read 309 drivers/net/wireless/marvell/mwifiex/uap_txrx.c if ((atomic_read(&adapter->pending_bridged_pkts) >= atomic_read 246 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&adapter->rx_pending) <= HIGH_RX_PENDING) { atomic_read 361 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&card->rx_cmd_urb_pending) && card->rx_cmd.urb) atomic_read 367 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&card->rx_data_urb_pending)) atomic_read 568 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&card->rx_cmd_urb_pending) && card->rx_cmd.urb) atomic_read 571 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&card->rx_data_urb_pending)) atomic_read 621 drivers/net/wireless/marvell/mwifiex/usb.c if (!atomic_read(&card->rx_data_urb_pending)) atomic_read 626 drivers/net/wireless/marvell/mwifiex/usb.c if (!atomic_read(&card->rx_cmd_urb_pending)) { atomic_read 839 drivers/net/wireless/marvell/mwifiex/usb.c atomic_read(&port->tx_data_urb_pending) == atomic_read 1080 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&port->tx_data_urb_pending) >= atomic_read 1141 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&port->tx_data_urb_pending) >= atomic_read 1193 drivers/net/wireless/marvell/mwifiex/usb.c if (atomic_read(&port->tx_data_urb_pending) atomic_read 1539 drivers/net/wireless/marvell/mwifiex/usb.c (!atomic_read(&card->rx_cmd_urb_pending))) atomic_read 700 drivers/net/wireless/marvell/mwifiex/util.c if (atomic_read(&phist_data->num_samples) > MWIFIEX_HIST_MAX_SAMPLES) atomic_read 485 drivers/net/wireless/marvell/mwifiex/wmm.c if (atomic_read(&priv->wmm.tx_pkts_queued)) atomic_read 662 drivers/net/wireless/marvell/mwifiex/wmm.c tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued); atomic_read 706 drivers/net/wireless/marvell/mwifiex/wmm.c tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued); atomic_read 887 drivers/net/wireless/marvell/mwifiex/wmm.c if (atomic_read(&priv->wmm.highest_queued_prio) < atomic_read 1106 drivers/net/wireless/marvell/mwifiex/wmm.c (atomic_read(&priv_tmp->wmm.tx_pkts_queued) == 0)) atomic_read 1115 drivers/net/wireless/marvell/mwifiex/wmm.c for (i = atomic_read(hqp); i >= LOW_PRIO_TID; --i) { atomic_read 1135 drivers/net/wireless/marvell/mwifiex/wmm.c if (atomic_read(&priv_tmp->wmm.tx_pkts_queued) != 0) { atomic_read 1152 drivers/net/wireless/marvell/mwifiex/wmm.c if (atomic_read(hqp) > i) atomic_read 1530 drivers/net/wireless/marvell/mwifiex/wmm.c if (atomic_read(&adapter->tx_queued) >= atomic_read 1560 drivers/net/wireless/marvell/mwl8k.c if (atomic_read(&priv->watchdog_event_pending)) atomic_read 1585 drivers/net/wireless/marvell/mwl8k.c if (atomic_read(&priv->watchdog_event_pending)) { atomic_read 71 drivers/net/wireless/mediatek/mt7601u/debugfs.c atomic_read(&dev->avg_ampdu_len)); atomic_read 26 drivers/net/wireless/quantenna/qtnfmac/util.h return atomic_read(&list->size); atomic_read 221 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c seqno = atomic_read(&intf->seqno); atomic_read 64 drivers/net/wireless/rsi/rsi_91x_coex.c } while (atomic_read(&coex_cb->coex_tx_thread.thread_done) == 0); atomic_read 260 drivers/net/wireless/rsi/rsi_91x_main.c } while (atomic_read(&common->tx_thread.thread_done) == 0); atomic_read 75 drivers/net/wireless/rsi/rsi_91x_sdio_ops.c if (atomic_read(&sdev->rx_thread.thread_done)) atomic_read 40 drivers/net/wireless/rsi/rsi_91x_usb_ops.c if (atomic_read(&dev->rx_thread.thread_done)) atomic_read 40 drivers/net/wireless/rsi/rsi_common.h (atomic_read(&event->event_condition) == 0)); atomic_read 43 drivers/net/wireless/rsi/rsi_common.h (atomic_read(&event->event_condition) == 0), atomic_read 138 drivers/net/wireless/st/cw1200/bh.c (CW1200_BH_SUSPENDED == atomic_read(&priv->bh_suspend)), atomic_read 153 drivers/net/wireless/st/cw1200/bh.c (CW1200_BH_RESUMED == atomic_read(&priv->bh_suspend)), atomic_read 436 drivers/net/wireless/st/cw1200/bh.c !atomic_read(&priv->recent_scan)) { atomic_read 450 drivers/net/wireless/st/cw1200/bh.c (atomic_read(&priv->bh_rx) == 0) && atomic_read 451 drivers/net/wireless/st/cw1200/bh.c (atomic_read(&priv->bh_tx) == 0)) atomic_read 461 drivers/net/wireless/st/cw1200/bh.c 0 : atomic_read(&priv->bh_suspend); atomic_read 511 drivers/net/wireless/st/cw1200/bh.c !atomic_read(&priv->recent_scan)) { atomic_read 530 drivers/net/wireless/st/cw1200/bh.c CW1200_BH_RESUME == atomic_read(&priv->bh_suspend)); atomic_read 232 drivers/net/wireless/st/cw1200/debug.c atomic_read(&priv->bh_term) ? "terminated" : "alive"); atomic_read 234 drivers/net/wireless/st/cw1200/debug.c atomic_read(&priv->bh_rx)); atomic_read 236 drivers/net/wireless/st/cw1200/debug.c atomic_read(&priv->bh_tx)); atomic_read 260 drivers/net/wireless/st/cw1200/debug.c atomic_read(&priv->tx_lock) ? "locked" : "unlocked"); atomic_read 261 drivers/net/wireless/st/cw1200/debug.c if (atomic_read(&priv->tx_lock)) atomic_read 263 drivers/net/wireless/st/cw1200/debug.c atomic_read(&priv->tx_lock)); atomic_read 284 drivers/net/wireless/st/cw1200/debug.c atomic_read(&priv->scan.in_progress) ? "active" : "idle"); atomic_read 149 drivers/net/wireless/st/cw1200/pm.c if (atomic_read(&priv->bh_rx)) { atomic_read 271 drivers/net/wireless/st/cw1200/pm.c if (atomic_read(&priv->bh_rx)) { atomic_read 159 drivers/net/wireless/st/cw1200/sta.c atomic_read(&priv->tx_lock), atomic_read 173 drivers/net/wireless/st/cw1200/sta.c if (!priv->vif->p2p && !atomic_read(&priv->tx_lock)) atomic_read 1243 drivers/net/wireless/st/cw1200/sta.c if (atomic_read(&priv->scan.in_progress)) { atomic_read 1387 drivers/net/wireless/st/cw1200/sta.c if (atomic_read(&priv->scan.in_progress)) { atomic_read 1186 drivers/net/wireless/st/cw1200/wsm.c BUG_ON(!atomic_read(&priv->tx_lock)); atomic_read 365 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (atomic_read(&intr->read_regs_enabled)) { atomic_read 392 drivers/net/wireless/zydas/zd1211rw/zd_usb.c } else if (atomic_read(&intr->read_regs_enabled)) { atomic_read 421 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (int_num == CR_INTERRUPT && atomic_read(&intr->read_regs_enabled)) atomic_read 464 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (hdr->id != USB_INT_ID_REGS && atomic_read(&intr->read_regs_enabled)) atomic_read 1027 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (!atomic_read(&tx->enabled)) { atomic_read 1090 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (!atomic_read(&tx->enabled) || !tx->watchdog_enabled) atomic_read 1488 drivers/net/xen-netback/netback.c !atomic_read(&queue->inflight_packets); atomic_read 2075 drivers/net/xen-netfront.c data[i] = atomic_read((atomic_t *)(np + xennet_stats[i].offset)); atomic_read 846 drivers/ntb/test/ntb_perf.c return likely(atomic_read(&pthr->perf->tsync) > 0) ? 0 : -EINTR; atomic_read 948 drivers/ntb/test/ntb_perf.c (atomic_read(&pthr->dma_sync) == 0 || atomic_read 949 drivers/ntb/test/ntb_perf.c atomic_read(&perf->tsync) < 0)); atomic_read 951 drivers/ntb/test/ntb_perf.c if (atomic_read(&perf->tsync) < 0) atomic_read 1074 drivers/ntb/test/ntb_perf.c atomic_read(&perf->tsync) <= 0); atomic_read 175 drivers/ntb/test/ntb_pingpong.c count = atomic_read(&pp->count); atomic_read 434 drivers/nvdimm/bus.c atomic_read(&nvdimm_bus->ioctl_active) == 0); atomic_read 973 drivers/nvdimm/bus.c if (atomic_read(&nvdimm->busy)) atomic_read 342 drivers/nvdimm/dimm_devs.c return sprintf(buf, "%s\n", atomic_read(&nvdimm->busy) atomic_read 534 drivers/nvdimm/security.c if (atomic_read(&nvdimm->busy)) { atomic_read 542 drivers/nvdimm/security.c if (atomic_read(&nvdimm->busy)) { atomic_read 455 drivers/nvme/host/fc.c if (atomic_read(&lport->act_rport_cnt) == 0) atomic_read 855 drivers/nvme/host/fc.c if (atomic_read(&rport->act_ctrl_cnt) == 0) atomic_read 769 drivers/nvme/target/fc.c (!atomic_read(&queue->connected) || atomic_read 1762 drivers/nvme/target/fc.c sqtail = atomic_read(&q->sqtail) % q->sqsize; atomic_read 170 drivers/oprofile/event_buffer.c wait_event_interruptible(buffer_wait, atomic_read(&buffer_ready)); atomic_read 176 drivers/oprofile/event_buffer.c if (!atomic_read(&buffer_ready)) atomic_read 184 drivers/oprofile/oprofilefs.c return oprofilefs_ulong_to_user(atomic_read(val), buf, count, offset); atomic_read 88 drivers/pci/ats.c if (atomic_read(&dev->ats_ref_cnt)) atomic_read 420 drivers/pci/hotplug/cpci_hotplug_core.c inserted, extracted, atomic_read(&extracting)); atomic_read 423 drivers/pci/hotplug/cpci_hotplug_core.c else if (!atomic_read(&extracting)) { atomic_read 453 drivers/pci/hotplug/cpci_hotplug_core.c } while (atomic_read(&extracting) && !kthread_should_stop()); atomic_read 485 drivers/pci/hotplug/cpci_hotplug_core.c } while (atomic_read(&extracting) && !kthread_should_stop()); atomic_read 378 drivers/pci/hotplug/pciehp_ctrl.c !atomic_read(&ctrl->pending_events) && atomic_read 412 drivers/pci/hotplug/pciehp_ctrl.c !atomic_read(&ctrl->pending_events) && atomic_read 684 drivers/pci/hotplug/pciehp_hpc.c atomic_read(&ctrl->pending_events)) atomic_read 313 drivers/pci/pci-sysfs.c return sprintf(buf, "%u\n", atomic_read(&pdev->enable_cnt)); atomic_read 1940 drivers/pci/pci.c dev_WARN_ONCE(&dev->dev, atomic_read(&dev->enable_cnt) <= 0, atomic_read 897 drivers/pci/pcie/aer.c if (atomic_read(&dev->enable_cnt) == 0) atomic_read 81 drivers/pci/switch/switchtec.c stuser->event_cnt = atomic_read(&stdev->event_cnt); atomic_read 559 drivers/pci/switch/switchtec.c if (stuser->event_cnt != atomic_read(&stdev->event_cnt)) atomic_read 700 drivers/pci/switch/switchtec.c stuser->event_cnt = atomic_read(&stdev->event_cnt); atomic_read 1325 drivers/pcmcia/ds.c if (atomic_read(&p_dev->socket->present) != 0) atomic_read 1346 drivers/perf/arm-cci.c if (atomic_read(active_events) == 0) atomic_read 305 drivers/phy/motorola/phy-cpcap-usb.c if (!atomic_read(&ddata->active)) atomic_read 3261 drivers/power/supply/ab8500_charger.c if (atomic_read(&di->current_stepping_sessions)) atomic_read 549 drivers/power/supply/cpcap-battery.c if (!atomic_read(&ddata->active)) atomic_read 481 drivers/power/supply/cpcap-charger.c if (!atomic_read(&ddata->active)) atomic_read 421 drivers/power/supply/power_supply_core.c if (atomic_read(&psy->use_cnt) >= 0 && atomic_read 776 drivers/power/supply/power_supply_core.c if (atomic_read(&psy->use_cnt) <= 0) { atomic_read 790 drivers/power/supply/power_supply_core.c if (atomic_read(&psy->use_cnt) <= 0 || !psy->desc->set_property) atomic_read 800 drivers/power/supply/power_supply_core.c if (atomic_read(&psy->use_cnt) <= 0 || atomic_read 810 drivers/power/supply/power_supply_core.c if (atomic_read(&psy->use_cnt) <= 0 || atomic_read 1890 drivers/rapidio/devices/rio_mport_cdev.c if (atomic_read(&chdev->active) == 0) atomic_read 2069 drivers/rapidio/devices/rio_mport_cdev.c if (atomic_read(&md->active) == 0) atomic_read 2043 drivers/rapidio/rio_cm.c if (atomic_read(&rdev->state) != RIO_DEVICE_SHUTDOWN) atomic_read 12 drivers/ras/debugfs.c return atomic_read(&trace_count); atomic_read 18 drivers/ras/debugfs.c return atomic_read(&trace_count); atomic_read 618 drivers/remoteproc/stm32_rproc.c if (atomic_read(&rproc->power) > 0) atomic_read 306 drivers/reset/core.c if (WARN_ON(atomic_read(&rstc->deassert_count) != 0)) atomic_read 353 drivers/reset/core.c if (WARN_ON(atomic_read(&rstc->triggered_count) != 0)) atomic_read 356 drivers/reset/core.c if (WARN_ON(atomic_read(&rstc->deassert_count) == 0)) atomic_read 412 drivers/reset/core.c if (WARN_ON(atomic_read(&rstc->triggered_count) != 0)) atomic_read 3662 drivers/s390/block/dasd.c open_count = atomic_read(&device->block->open_count); atomic_read 2266 drivers/s390/block/dasd_3990_erp.c if (atomic_read(&device->path[pos].error_count) >= atomic_read 622 drivers/s390/block/dasd_devmap.c wait_event(dasd_delete_wq, atomic_read(&device->ref_count) == 0); atomic_read 502 drivers/s390/block/dasd_ioctl.c dasd_info->open_count = atomic_read(&block->open_count); atomic_read 355 drivers/s390/block/dcssblk.c if (atomic_read(&dev_info->use_count)) { atomic_read 462 drivers/s390/block/dcssblk.c if (atomic_read(&dev_info->use_count) == 0) { atomic_read 777 drivers/s390/block/dcssblk.c if (atomic_read(&dev_info->use_count) != 0) { atomic_read 213 drivers/s390/char/monreader.c if (!atomic_read(&monpriv->read_ready)) atomic_read 313 drivers/s390/char/monreader.c atomic_read(&monpriv->iucv_connected) || atomic_read 314 drivers/s390/char/monreader.c atomic_read(&monpriv->iucv_severed)); atomic_read 315 drivers/s390/char/monreader.c if (atomic_read(&monpriv->iucv_severed)) { atomic_read 382 drivers/s390/char/monreader.c atomic_read(&monpriv->read_ready) || atomic_read 383 drivers/s390/char/monreader.c atomic_read(&monpriv->iucv_severed)); atomic_read 386 drivers/s390/char/monreader.c if (unlikely(atomic_read(&monpriv->iucv_severed))) atomic_read 437 drivers/s390/char/monreader.c if (unlikely(atomic_read(&monpriv->iucv_severed))) atomic_read 439 drivers/s390/char/monreader.c if (atomic_read(&monpriv->read_ready)) atomic_read 506 drivers/s390/char/monreader.c atomic_read(&monpriv->iucv_connected) || atomic_read 507 drivers/s390/char/monreader.c atomic_read(&monpriv->iucv_severed)); atomic_read 508 drivers/s390/char/monreader.c if (atomic_read(&monpriv->iucv_severed)) atomic_read 1015 drivers/s390/char/raw3270.c wait_event(raw3270_wait_queue, atomic_read(&view->ref_count) == 0); atomic_read 417 drivers/s390/char/vmlogrdr.c if (atomic_read(&priv->receive_ready)) { atomic_read 486 drivers/s390/char/vmlogrdr.c atomic_read(&priv->receive_ready)); atomic_read 494 drivers/s390/cio/ccwgroup.c if (atomic_read(&gdev->onoff)) atomic_read 71 drivers/s390/cio/crw.c atomic_read(&crw_nr_req) > 0); atomic_read 146 drivers/s390/cio/crw.c wait_event(crw_handler_wait_q, atomic_read(&crw_nr_req) == 0); atomic_read 1328 drivers/s390/cio/css.c atomic_read(&css_eval_scheduled) == 0); atomic_read 170 drivers/s390/cio/device.c atomic_read(&ccw_device_init_count) == 0); atomic_read 1808 drivers/s390/cio/device.c if (atomic_read(&cdev->private->onoff)) atomic_read 125 drivers/s390/cio/qdio_debug.c atomic_read(&q->nr_buf_used), q->first_to_check); atomic_read 511 drivers/s390/cio/qdio_main.c count = min(atomic_read(&q->nr_buf_used), QDIO_MAX_BUFFERS_MASK); atomic_read 568 drivers/s390/cio/qdio_main.c if (!atomic_read(&q->nr_buf_used)) atomic_read 722 drivers/s390/cio/qdio_main.c count = atomic_read(&q->nr_buf_used); atomic_read 766 drivers/s390/cio/qdio_main.c return atomic_read(&q->nr_buf_used) == 0; atomic_read 832 drivers/s390/cio/qdio_main.c WARN_ON_ONCE(atomic_read(&q->nr_buf_used) < 0); atomic_read 138 drivers/s390/cio/qdio_thinint.c if (!atomic_read(&q_indicators[TIQDIO_SHARED_IND].count)) atomic_read 92 drivers/s390/cio/vfio_ccw_ops.c return sprintf(buf, "%d\n", atomic_read(&private->avail)); atomic_read 379 drivers/s390/crypto/vfio_ap_ops.c atomic_read(&matrix_dev->available_instances)); atomic_read 80 drivers/s390/crypto/zcrypt_api.c if (atomic_read(&zcrypt_rescan_req)) { atomic_read 605 drivers/s390/crypto/zcrypt_api.c weight += atomic_read(&zc->load); atomic_read 606 drivers/s390/crypto/zcrypt_api.c pref_weight += atomic_read(&pref_zc->load); atomic_read 620 drivers/s390/crypto/zcrypt_api.c weight += atomic_read(&zq->load); atomic_read 621 drivers/s390/crypto/zcrypt_api.c pref_weight += atomic_read(&pref_zq->load); atomic_read 1432 drivers/s390/crypto/zcrypt_api.c return put_user(atomic_read(&zcrypt_open_count), atomic_read 89 drivers/s390/crypto/zcrypt_card.c return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&zc->load)); atomic_read 81 drivers/s390/crypto/zcrypt_queue.c return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&zq->load)); atomic_read 126 drivers/s390/net/fsm.c int st = atomic_read(&fi->state); atomic_read 147 drivers/s390/net/fsm.h int state = atomic_read(&fi->state); atomic_read 215 drivers/s390/net/fsm.h return atomic_read(&fi->state); atomic_read 553 drivers/s390/net/qeth_core.h return atomic_read(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q; atomic_read 558 drivers/s390/net/qeth_core.h return atomic_read(&queue->used_buffers) == 0; atomic_read 407 drivers/s390/net/qeth_core_main.c atomic_read(&c->state) == atomic_read 428 drivers/s390/net/qeth_core_main.c if (forced_cleanup && (atomic_read(&(q->bufs[bidx]->state)) == atomic_read 455 drivers/s390/net/qeth_core_main.c WARN_ON_ONCE(atomic_read(&buffer->state) != atomic_read 1098 drivers/s390/net/qeth_core_main.c WARN_ON_ONCE(atomic_read(&buf->state) == QETH_QDIO_BUF_IN_CQ); atomic_read 1100 drivers/s390/net/qeth_core_main.c if (atomic_read(&buf->state) == QETH_QDIO_BUF_PENDING) atomic_read 1241 drivers/s390/net/qeth_core_main.c if (atomic_read(&card->qdio.state) != QETH_QDIO_UNINITIALIZED) atomic_read 3227 drivers/s390/net/qeth_core_main.c if ((atomic_read(&buffer->state) == QETH_QDIO_BUF_EMPTY) && atomic_read 3245 drivers/s390/net/qeth_core_main.c if (atomic_read(&queue->used_buffers) atomic_read 3264 drivers/s390/net/qeth_core_main.c if (atomic_read(&queue->used_buffers) atomic_read 3298 drivers/s390/net/qeth_core_main.c if ((atomic_read(&queue->used_buffers) >= atomic_read 3301 drivers/s390/net/qeth_core_main.c !atomic_read(&queue->set_pci_flags_count)) { atomic_read 3308 drivers/s390/net/qeth_core_main.c if (!atomic_read(&queue->set_pci_flags_count)) { atomic_read 3324 drivers/s390/net/qeth_core_main.c if (atomic_read(&queue->set_pci_flags_count)) atomic_read 3368 drivers/s390/net/qeth_core_main.c if ((atomic_read(&queue->used_buffers) <= QETH_LOW_WATERMARK_PACK) || atomic_read 3369 drivers/s390/net/qeth_core_main.c !atomic_read(&queue->set_pci_flags_count)) { atomic_read 3383 drivers/s390/net/qeth_core_main.c !atomic_read(&queue->set_pci_flags_count)) atomic_read 3844 drivers/s390/net/qeth_core_main.c if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) atomic_read 3854 drivers/s390/net/qeth_core_main.c if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) atomic_read 3909 drivers/s390/net/qeth_core_main.c if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) { atomic_read 3932 drivers/s390/net/qeth_core_main.c if (atomic_read(&buffer->state) != atomic_read 3967 drivers/s390/net/qeth_core_main.c else if (!atomic_read(&queue->set_pci_flags_count)) atomic_read 3983 drivers/s390/net/qeth_core_main.c if (!tmp && !atomic_read(&queue->set_pci_flags_count)) atomic_read 5087 drivers/s390/net/qeth_core_main.c (!atomic_read(&card->force_alloc_skb))) || atomic_read 277 drivers/s390/scsi/zfcp_aux.c if (atomic_read(&adapter->stat_miss) >= atomic_read 250 drivers/s390/scsi/zfcp_ccw.c if (atomic_read(&adapter->status) & atomic_read 264 drivers/s390/scsi/zfcp_ccw.c if (atomic_read(&adapter->status) & atomic_read 254 drivers/s390/scsi/zfcp_dbf.c rec->adapter_status = atomic_read(&adapter->status); atomic_read 256 drivers/s390/scsi/zfcp_dbf.c rec->port_status = atomic_read(&port->status); atomic_read 261 drivers/s390/scsi/zfcp_dbf.c rec->lun_status = atomic_read(&sdev_to_zfcp(sdev)->status); atomic_read 364 drivers/s390/scsi/zfcp_dbf.c atomic_read(&sdev_to_zfcp(erp->sdev)->erp_counter); atomic_read 366 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_count = atomic_read(&erp->port->erp_counter); atomic_read 368 drivers/s390/scsi/zfcp_dbf.c rec->u.run.rec_count = atomic_read(&erp->adapter->erp_counter); atomic_read 339 drivers/s390/scsi/zfcp_def.h return atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_MB_ACT; atomic_read 87 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_ERP_INUSE) atomic_read 95 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_INUSE) atomic_read 110 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_INUSE) atomic_read 130 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zsdev->status) & ZFCP_STATUS_COMMON_ERP_FAILED) atomic_read 134 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED) atomic_read 138 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&port->status) & atomic_read 147 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&adapter->status) & atomic_read 172 drivers/s390/scsi/zfcp_erp.c l_status = atomic_read(&zfcp_sdev->status); atomic_read 175 drivers/s390/scsi/zfcp_erp.c p_status = atomic_read(&port->status); atomic_read 183 drivers/s390/scsi/zfcp_erp.c p_status = atomic_read(&port->status); atomic_read 188 drivers/s390/scsi/zfcp_erp.c p_status = atomic_read(&port->status); atomic_read 191 drivers/s390/scsi/zfcp_erp.c a_status = atomic_read(&adapter->status); atomic_read 201 drivers/s390/scsi/zfcp_erp.c a_status = atomic_read(&adapter->status); atomic_read 238 drivers/s390/scsi/zfcp_erp.c if (!(atomic_read(&zfcp_sdev->status) & atomic_read 252 drivers/s390/scsi/zfcp_erp.c if (!(atomic_read(&port->status) & ZFCP_STATUS_COMMON_RUNNING)) atomic_read 263 drivers/s390/scsi/zfcp_erp.c if (!(atomic_read(&adapter->status) & atomic_read 529 drivers/s390/scsi/zfcp_erp.c return (atomic_read(status) ^ mask) & mask; atomic_read 757 drivers/s390/scsi/zfcp_erp.c if (!(atomic_read(&adapter->status) & atomic_read 768 drivers/s390/scsi/zfcp_erp.c if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_XCONFIG_OK)) atomic_read 870 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_OPEN) { atomic_read 903 drivers/s390/scsi/zfcp_erp.c int status = atomic_read(&port->status); atomic_read 973 drivers/s390/scsi/zfcp_erp.c int p_status = atomic_read(&port->status); atomic_read 1014 drivers/s390/scsi/zfcp_erp.c int p_status = atomic_read(&port->status); atomic_read 1086 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_OPEN) atomic_read 1091 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_OPEN) atomic_read 1098 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_OPEN) atomic_read 1122 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zfcp_sdev->erp_counter) > ZFCP_MAX_ERPS) { atomic_read 1140 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_ERP_FAILED) { atomic_read 1157 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_NOESC) { atomic_read 1162 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&port->erp_counter) > ZFCP_MAX_ERPS) { atomic_read 1178 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED) { atomic_read 1196 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&adapter->erp_counter) > ZFCP_MAX_ERPS) { atomic_read 1212 drivers/s390/scsi/zfcp_erp.c if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_FAILED) { atomic_read 1246 drivers/s390/scsi/zfcp_erp.c int status = atomic_read(target_status); atomic_read 1349 drivers/s390/scsi/zfcp_erp.c port_status = atomic_read(&port->status); atomic_read 1371 drivers/s390/scsi/zfcp_erp.c lun_status = atomic_read(&zsdev->status); atomic_read 1611 drivers/s390/scsi/zfcp_erp.c !(atomic_read(&adapter->status) & atomic_read 178 drivers/s390/scsi/zfcp_fc.c if ((atomic_read(&wka_port->refcount) != 0) || atomic_read 518 drivers/s390/scsi/zfcp_fc.c !(atomic_read(&port->status) & ZFCP_STATUS_COMMON_OPEN)) { atomic_read 581 drivers/s390/scsi/zfcp_fc.c if (atomic_read(&port->status) & ZFCP_STATUS_PORT_LINK_TEST) atomic_read 699 drivers/s390/scsi/zfcp_fc.c if (!(atomic_read(&port->status) & ZFCP_STATUS_COMMON_NOESC)) atomic_read 1057 drivers/s390/scsi/zfcp_fc.c if (!(atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_OPEN)) atomic_read 126 drivers/s390/scsi/zfcp_fsf.c if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED) atomic_read 440 drivers/s390/scsi/zfcp_fsf.c BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP); atomic_read 762 drivers/s390/scsi/zfcp_fsf.c req->qdio_req.qdio_outb_usage = atomic_read(&qdio->req_q_free); atomic_read 929 drivers/s390/scsi/zfcp_fsf.c if (unlikely(!(atomic_read(&zfcp_sdev->status) & atomic_read 2282 drivers/s390/scsi/zfcp_fsf.c if (unlikely(!(atomic_read(&zfcp_sdev->status) & atomic_read 2287 drivers/s390/scsi/zfcp_fsf.c if (atomic_read(&qdio->req_q_free) <= 0) { atomic_read 2393 drivers/s390/scsi/zfcp_fsf.c if (unlikely(!(atomic_read(&zfcp_sdev->status) & atomic_read 57 drivers/s390/scsi/zfcp_qdio.c used = QDIO_MAX_BUFFERS_PER_Q - atomic_read(&qdio->req_q_free); atomic_read 210 drivers/s390/scsi/zfcp_qdio.c if (atomic_read(&qdio->req_q_free) || atomic_read 211 drivers/s390/scsi/zfcp_qdio.c !(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) atomic_read 233 drivers/s390/scsi/zfcp_qdio.c if (!(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) atomic_read 347 drivers/s390/scsi/zfcp_qdio.c if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) atomic_read 360 drivers/s390/scsi/zfcp_qdio.c count = atomic_read(&qdio->req_q_free); atomic_read 384 drivers/s390/scsi/zfcp_qdio.c if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP) atomic_read 497 drivers/s390/scsi/zfcp_qdio.c if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_SIOSL_ISSUED) atomic_read 114 drivers/s390/scsi/zfcp_qdio.h int count = min(atomic_read(&qdio->req_q_free), atomic_read 209 drivers/s390/scsi/zfcp_qdio.h int count = min(atomic_read(&qdio->req_q_free), max_sbals); atomic_read 85 drivers/s390/scsi/zfcp_scsi.c status = atomic_read(&zfcp_sdev->status); atomic_read 87 drivers/s390/scsi/zfcp_scsi.c !(atomic_read(&zfcp_sdev->port->status) & atomic_read 204 drivers/s390/scsi/zfcp_scsi.c if (!(atomic_read(&adapter->status) & atomic_read 311 drivers/s390/scsi/zfcp_scsi.c if (!(atomic_read(&adapter->status) & atomic_read 651 drivers/s390/scsi/zfcp_scsi.c int status = atomic_read(&adapter->status); atomic_read 808 drivers/s390/scsi/zfcp_scsi.c data_div = atomic_read(&adapter->status) & atomic_read 60 drivers/s390/scsi/zfcp_sysfs.c ZFCP_DEFINE_A_ATTR(status, "0x%08x\n", atomic_read(&adapter->status)); atomic_read 69 drivers/s390/scsi/zfcp_sysfs.c ZFCP_DEFINE_A_ATTR(in_recovery, "%d\n", (atomic_read(&adapter->status) & atomic_read 73 drivers/s390/scsi/zfcp_sysfs.c atomic_read(&port->status)); atomic_read 75 drivers/s390/scsi/zfcp_sysfs.c (atomic_read(&port->status) & atomic_read 96 drivers/s390/scsi/zfcp_sysfs.c if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED) atomic_read 132 drivers/s390/scsi/zfcp_sysfs.c status = atomic_read(&sdev_to_zfcp(sdev)->status); atomic_read 177 drivers/s390/scsi/zfcp_sysfs.c if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_FAILED) atomic_read 247 drivers/s390/scsi/zfcp_sysfs.c return atomic_read(&port->units) == -1; atomic_read 258 drivers/s390/scsi/zfcp_sysfs.c if (atomic_read(&port->units) > 0) atomic_read 506 drivers/s390/scsi/zfcp_sysfs.c (atomic_read(&zfcp_sdev->status) & atomic_read 514 drivers/s390/scsi/zfcp_sysfs.c unsigned int status = atomic_read(&sdev_to_zfcp(sdev)->status); atomic_read 542 drivers/s390/scsi/zfcp_sysfs.c (atomic_read(&zfcp_sdev->status) & atomic_read 546 drivers/s390/scsi/zfcp_sysfs.c atomic_read(&zfcp_sdev->status)); atomic_read 654 drivers/s390/scsi/zfcp_sysfs.c return sprintf(buf, "%d %llu\n", atomic_read(&qdio->req_q_full), atomic_read 223 drivers/s390/scsi/zfcp_unit.c status = atomic_read(&zfcp_sdev->status); atomic_read 395 drivers/scsi/aacraid/commsup.c qid, atomic_read(&q->numpending)); atomic_read 277 drivers/scsi/arcmsr/arcmsr_attr.c atomic_read(&acb->ccboutstandingcount)); atomic_read 1402 drivers/scsi/arcmsr/arcmsr_hba.c , atomic_read(&acb->ccboutstandingcount)); atomic_read 1482 drivers/scsi/arcmsr/arcmsr_hba.c residual = atomic_read(&acb->ccboutstandingcount); atomic_read 1615 drivers/scsi/arcmsr/arcmsr_hba.c if (!atomic_read(&acb->ccboutstandingcount)) atomic_read 1621 drivers/scsi/arcmsr/arcmsr_hba.c if (atomic_read(&acb->ccboutstandingcount)) { atomic_read 3357 drivers/scsi/arcmsr/arcmsr_hba.c , atomic_read(&acb->ccboutstandingcount)); atomic_read 3426 drivers/scsi/arcmsr/arcmsr_hba.c , atomic_read(&acb->ccboutstandingcount)); atomic_read 3488 drivers/scsi/arcmsr/arcmsr_hba.c , atomic_read(&acb->ccboutstandingcount)); atomic_read 3565 drivers/scsi/arcmsr/arcmsr_hba.c , atomic_read(&acb->ccboutstandingcount)); atomic_read 3633 drivers/scsi/arcmsr/arcmsr_hba.c , atomic_read(&acb->ccboutstandingcount)); atomic_read 3968 drivers/scsi/arcmsr/arcmsr_hba.c if (unlikely(atomic_read(&acb->rq_map_token) == 0) || atomic_read 3975 drivers/scsi/arcmsr/arcmsr_hba.c if (atomic_read(&acb->ante_token_value) == atomic_read 3976 drivers/scsi/arcmsr/arcmsr_hba.c atomic_read(&acb->rq_map_token)) { atomic_read 3980 drivers/scsi/arcmsr/arcmsr_hba.c atomic_read(&acb->rq_map_token)); atomic_read 4341 drivers/scsi/arcmsr/arcmsr_hba.c if (atomic_read(&acb->ccboutstandingcount) != 0) { atomic_read 4454 drivers/scsi/arcmsr/arcmsr_hba.c if (!atomic_read(&acb->ccboutstandingcount)) { atomic_read 500 drivers/scsi/bfa/bfad_debugfs.c if (atomic_read(&bfa_debugfs_port_count) == 0) { atomic_read 424 drivers/scsi/bnx2fc/bnx2fc_io.c free_sqes = atomic_read(&tgt->free_sqes); atomic_read 483 drivers/scsi/bnx2fc/bnx2fc_io.c free_sqes = atomic_read(&tgt->free_sqes); atomic_read 160 drivers/scsi/bnx2i/bnx2i_hwi.c num_active_cmds = atomic_read(&ep->num_active_cmds); atomic_read 2045 drivers/scsi/bnx2i/bnx2i_hwi.c if (!atomic_read(&bnx2i_conn->ep->num_active_cmds)) atomic_read 1234 drivers/scsi/bnx2i/bnx2i_iscsi.c if (atomic_read(&bnx2i_conn->ep->num_active_cmds) + 1 > atomic_read 1490 drivers/scsi/bnx2i/bnx2i_iscsi.c if (atomic_read(&bnx2i_conn->work_cnt)) { atomic_read 706 drivers/scsi/cxlflash/main.c while (atomic_read(&afu->cmds_active)) atomic_read 1705 drivers/scsi/cxlflash/superpipe.c ((atomic_read(&cfg->recovery_threads) > 1) || atomic_read 83 drivers/scsi/esas2r/esas2r_int.c if (atomic_read(&a->disable_cnt) == 0) atomic_read 127 drivers/scsi/esas2r/esas2r_int.c if (likely(atomic_read(&a->disable_cnt) == 0)) atomic_read 162 drivers/scsi/esas2r/esas2r_int.c if (likely(atomic_read(&a->disable_cnt) == 0)) atomic_read 471 drivers/scsi/esas2r/esas2r_int.c if (atomic_read(&a->disable_cnt) == 0) atomic_read 661 drivers/scsi/esas2r/esas2r_int.c if (atomic_read(&a->disable_cnt) == 0) atomic_read 754 drivers/scsi/esas2r/esas2r_io.c if (atomic_read(&a->disable_cnt) == 0) atomic_read 835 drivers/scsi/esas2r/esas2r_io.c if (atomic_read(&a->disable_cnt) == 0) atomic_read 1031 drivers/scsi/esas2r/esas2r_main.c if (atomic_read(&a->disable_cnt) == 0) atomic_read 218 drivers/scsi/fnic/fnic_scsi.c while (atomic_read(&fnic->in_flight)) atomic_read 563 drivers/scsi/hosts.c return atomic_read(&shost->host_busy); atomic_read 529 drivers/scsi/hpsa.c atomic_read(&h->commands_outstanding)); atomic_read 2453 drivers/scsi/hpsa.c atomic_read(&dev->commands_outstanding) <= 0) atomic_read 3169 drivers/scsi/hpsa.c atomic_read(&dev->commands_outstanding) <= 0 || atomic_read 6039 drivers/scsi/hpsa.c if (atomic_read(&dev->commands_outstanding) > 0) atomic_read 1082 drivers/scsi/hptiop.c atomic_read(&hba->resetting) == 0, 60 * HZ); atomic_read 1084 drivers/scsi/hptiop.c if (atomic_read(&hba->resetting)) { atomic_read 1243 drivers/scsi/ibmvscsi/ibmvfc.c BUG_ON(atomic_read(&pool->events[i].free) != 1); atomic_read 2781 drivers/scsi/ibmvscsi/ibmvfc.c if (unlikely(atomic_read(&evt->free))) { atomic_read 490 drivers/scsi/ibmvscsi/ibmvscsi.c if (atomic_read(&pool->events[i].free) != 1) atomic_read 1730 drivers/scsi/ibmvscsi/ibmvscsi.c atomic_read(&hostdata->request_limit) < 2;) { atomic_read 1735 drivers/scsi/ibmvscsi/ibmvscsi.c if (atomic_read(&hostdata->request_limit) <= 0) atomic_read 1814 drivers/scsi/ibmvscsi/ibmvscsi.c if (atomic_read(&evt_struct->free)) { atomic_read 2288 drivers/scsi/ibmvscsi/ibmvscsi.c atomic_read(&hostdata->request_limit) < 2;) { atomic_read 2294 drivers/scsi/ibmvscsi/ibmvscsi.c if (atomic_read(&hostdata->request_limit) > 0) atomic_read 117 drivers/scsi/iscsi_tcp.c !atomic_read(&sk->sk_rmem_alloc)) { atomic_read 2355 drivers/scsi/libfc/fc_exch.c st->fc_no_free_exch += atomic_read(&mp->stats.no_free_exch); atomic_read 2357 drivers/scsi/libfc/fc_exch.c atomic_read(&mp->stats.no_free_exch_xid); atomic_read 2358 drivers/scsi/libfc/fc_exch.c st->fc_xid_not_found += atomic_read(&mp->stats.xid_not_found); atomic_read 2359 drivers/scsi/libfc/fc_exch.c st->fc_xid_busy += atomic_read(&mp->stats.xid_busy); atomic_read 2360 drivers/scsi/libfc/fc_exch.c st->fc_seq_not_found += atomic_read(&mp->stats.seq_not_found); atomic_read 2361 drivers/scsi/libfc/fc_exch.c st->fc_non_bls_resp += atomic_read(&mp->stats.non_bls_resp); atomic_read 608 drivers/scsi/libsas/sas_init.c if (atomic_read(&phy->event_nr) > phy->ha->event_thres) { atomic_read 220 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_ls_req_in), atomic_read 221 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_ls_req_drop), atomic_read 222 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_abort)); atomic_read 226 drivers/scsi/lpfc/lpfc_attr.c if (atomic_read(&tgtp->rcv_ls_req_in) != atomic_read 227 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_ls_req_out)) { atomic_read 230 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_ls_req_in), atomic_read 231 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_ls_req_out)); atomic_read 238 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_rsp), atomic_read 239 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_drop), atomic_read 240 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_rsp_cmpl)); atomic_read 246 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_rsp_aborted), atomic_read 247 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_rsp_xb_set), atomic_read 248 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_ls_rsp_error)); atomic_read 255 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 256 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_fcp_cmd_defer), atomic_read 257 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_release), atomic_read 258 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_fcp_cmd_drop)); atomic_read 262 drivers/scsi/lpfc/lpfc_attr.c if (atomic_read(&tgtp->rcv_fcp_cmd_in) != atomic_read 263 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_fcp_cmd_out)) { atomic_read 266 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 267 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->rcv_fcp_cmd_out)); atomic_read 275 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_read), atomic_read 276 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_read_rsp), atomic_read 277 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_write), atomic_read 278 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_rsp), atomic_read 279 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_drop)); atomic_read 285 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_rsp_cmpl), atomic_read 286 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_rsp_error), atomic_read 287 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_rsp_drop)); atomic_read 293 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_rsp_aborted), atomic_read 294 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_rsp_xb_set), atomic_read 295 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_xri_abort_cqe)); atomic_read 301 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_abort), atomic_read 302 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_fcp_abort_cmpl)); atomic_read 308 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_abort_sol), atomic_read 309 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_abort_unsol), atomic_read 310 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_abort_rsp), atomic_read 311 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->xmt_abort_rsp_error)); atomic_read 317 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->defer_ctx), atomic_read 318 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->defer_fod), atomic_read 319 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&tgtp->defer_wqfull)); atomic_read 324 drivers/scsi/lpfc/lpfc_attr.c tot = atomic_read(&tgtp->rcv_fcp_cmd_drop); atomic_read 325 drivers/scsi/lpfc/lpfc_attr.c tot += atomic_read(&tgtp->xmt_fcp_release); atomic_read 326 drivers/scsi/lpfc/lpfc_attr.c tot = atomic_read(&tgtp->rcv_fcp_cmd_in) - tot; atomic_read 458 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->fc4NvmeLsRequests), atomic_read 459 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->fc4NvmeLsCmpls), atomic_read 460 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_ls_abort)); atomic_read 466 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_ls_err), atomic_read 467 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->cmpl_ls_xb), atomic_read 468 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->cmpl_ls_err)); atomic_read 493 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_fcp_abort), atomic_read 494 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_fcp_noxri), atomic_read 495 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_fcp_bad_ndlp), atomic_read 496 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_fcp_qdepth), atomic_read 497 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_fcp_err), atomic_read 498 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->xmt_fcp_wqerr)); atomic_read 504 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->cmpl_fcp_xb), atomic_read 505 drivers/scsi/lpfc/lpfc_attr.c atomic_read(&lport->cmpl_fcp_err)); atomic_read 163 drivers/scsi/lpfc/lpfc_debugfs.c index = (atomic_read(&vport->disc_trc_cnt) + 1) & atomic_read 229 drivers/scsi/lpfc/lpfc_debugfs.c index = (atomic_read(&phba->slow_ring_trc_cnt) + 1) & atomic_read 905 drivers/scsi/lpfc/lpfc_debugfs.c i = atomic_read(&ndlp->cmd_pending); atomic_read 1055 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_ls_req_in), atomic_read 1056 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_ls_req_drop), atomic_read 1057 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_abort)); atomic_read 1058 drivers/scsi/lpfc/lpfc_debugfs.c if (atomic_read(&tgtp->rcv_ls_req_in) != atomic_read 1059 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_ls_req_out)) { atomic_read 1062 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_ls_req_in), atomic_read 1063 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_ls_req_out)); atomic_read 1068 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_rsp), atomic_read 1069 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_drop), atomic_read 1070 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_rsp_cmpl)); atomic_read 1074 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_rsp_aborted), atomic_read 1075 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_rsp_xb_set), atomic_read 1076 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_ls_rsp_error)); atomic_read 1081 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 1082 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_fcp_cmd_defer), atomic_read 1083 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_release), atomic_read 1084 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_fcp_cmd_drop)); atomic_read 1086 drivers/scsi/lpfc/lpfc_debugfs.c if (atomic_read(&tgtp->rcv_fcp_cmd_in) != atomic_read 1087 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_fcp_cmd_out)) { atomic_read 1090 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 1091 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->rcv_fcp_cmd_out)); atomic_read 1097 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_read), atomic_read 1098 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_read_rsp), atomic_read 1099 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_write), atomic_read 1100 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_rsp)); atomic_read 1104 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_rsp_cmpl), atomic_read 1105 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_rsp_error), atomic_read 1106 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_rsp_drop)); atomic_read 1110 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_rsp_aborted), atomic_read 1111 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_rsp_xb_set), atomic_read 1112 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_xri_abort_cqe)); atomic_read 1116 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_abort), atomic_read 1117 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_fcp_abort_cmpl)); atomic_read 1121 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_abort_sol), atomic_read 1122 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_abort_unsol), atomic_read 1123 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_abort_rsp), atomic_read 1124 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&tgtp->xmt_abort_rsp_error)); atomic_read 1155 drivers/scsi/lpfc/lpfc_debugfs.c tot = atomic_read(&tgtp->rcv_fcp_cmd_drop); atomic_read 1156 drivers/scsi/lpfc/lpfc_debugfs.c tot += atomic_read(&tgtp->xmt_fcp_release); atomic_read 1157 drivers/scsi/lpfc/lpfc_debugfs.c tot = atomic_read(&tgtp->rcv_fcp_cmd_in) - tot; atomic_read 1182 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->fc4NvmeLsRequests), atomic_read 1183 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->fc4NvmeLsCmpls)); atomic_read 1216 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_ls_abort), atomic_read 1217 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_ls_err), atomic_read 1218 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->cmpl_ls_xb), atomic_read 1219 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->cmpl_ls_err)); atomic_read 1224 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_fcp_noxri), atomic_read 1225 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_fcp_bad_ndlp), atomic_read 1226 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_fcp_qdepth), atomic_read 1227 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_fcp_wqerr), atomic_read 1228 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_fcp_err), atomic_read 1229 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->xmt_fcp_abort)); atomic_read 1233 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->cmpl_fcp_xb), atomic_read 1234 drivers/scsi/lpfc/lpfc_debugfs.c atomic_read(&lport->cmpl_fcp_err)); atomic_read 1524 drivers/scsi/lpfc/lpfc_debugfs.c index = (atomic_read(&phba->nvmeio_trc_cnt) + 1) & atomic_read 6112 drivers/scsi/lpfc/lpfc_debugfs.c if (atomic_read(&phba->debugfs_vport_count) == 0) { atomic_read 6217 drivers/scsi/lpfc/lpfc_debugfs.c if (atomic_read(&lpfc_debugfs_hba_count) == 0) { atomic_read 9610 drivers/scsi/lpfc/lpfc_els.c if (atomic_read(&phba->fabric_iocb_count) == 0) { atomic_read 9730 drivers/scsi/lpfc/lpfc_els.c BUG_ON(atomic_read(&phba->fabric_iocb_count) == 0); atomic_read 9775 drivers/scsi/lpfc/lpfc_els.c BUG_ON(atomic_read(&phba->fabric_iocb_count) > 1); atomic_read 9778 drivers/scsi/lpfc/lpfc_els.c ready = atomic_read(&phba->fabric_iocb_count) == 0 && atomic_read 412 drivers/scsi/lpfc/lpfc_hbadisc.c if (atomic_read(&phba->fast_event_count) > LPFC_MAX_EVT_COUNT) atomic_read 1611 drivers/scsi/lpfc/lpfc_nvme.c if ((atomic_read(&ndlp->cmd_pending) >= ndlp->cmd_qdepth) && atomic_read 1617 drivers/scsi/lpfc/lpfc_nvme.c atomic_read(&ndlp->cmd_pending), atomic_read 474 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 475 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->rcv_fcp_cmd_out), atomic_read 476 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->xmt_fcp_release)); atomic_read 2133 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 2134 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->rcv_fcp_cmd_out), atomic_read 2135 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->xmt_fcp_release)); atomic_read 2371 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 2372 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->rcv_fcp_cmd_out), atomic_read 2373 drivers/scsi/lpfc/lpfc_nvmet.c atomic_read(&tgtp->xmt_fcp_release)); atomic_read 252 drivers/scsi/lpfc/lpfc_scsi.c num_rsrc_err = atomic_read(&phba->num_rsrc_err); atomic_read 253 drivers/scsi/lpfc/lpfc_scsi.c num_cmd_success = atomic_read(&phba->num_cmd_success); atomic_read 4017 drivers/scsi/lpfc/lpfc_scsi.c atomic_read(&pnode->cmd_pending) && atomic_read 4018 drivers/scsi/lpfc/lpfc_scsi.c (atomic_read(&pnode->cmd_pending) > atomic_read 4023 drivers/scsi/lpfc/lpfc_scsi.c atomic_read(&pnode->cmd_pending); atomic_read 4542 drivers/scsi/lpfc/lpfc_scsi.c if (atomic_read(&ndlp->cmd_pending) >= ndlp->cmd_qdepth) { atomic_read 4551 drivers/scsi/lpfc/lpfc_scsi.c atomic_read(&ndlp->cmd_pending), atomic_read 13443 drivers/scsi/lpfc/lpfc_sli.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 13444 drivers/scsi/lpfc/lpfc_sli.c atomic_read(&tgtp->rcv_fcp_cmd_out), atomic_read 13445 drivers/scsi/lpfc/lpfc_sli.c atomic_read(&tgtp->xmt_fcp_release)); atomic_read 13953 drivers/scsi/lpfc/lpfc_sli.c atomic_read(&tgtp->rcv_fcp_cmd_in), atomic_read 13954 drivers/scsi/lpfc/lpfc_sli.c atomic_read(&tgtp->rcv_fcp_cmd_out), atomic_read 13955 drivers/scsi/lpfc/lpfc_sli.c atomic_read(&tgtp->xmt_fcp_release)); atomic_read 403 drivers/scsi/megaraid.c if (atomic_read(&adapter->quiescent) == 0) atomic_read 1310 drivers/scsi/megaraid.c if(atomic_read(&adapter->quiescent) == 0) { atomic_read 1390 drivers/scsi/megaraid.c if(atomic_read(&adapter->quiescent) == 0) { atomic_read 2101 drivers/scsi/megaraid.c atomic_read(&adapter->quiescent)); atomic_read 2126 drivers/scsi/megaraid.c seq_printf(m, "pend_cmds = %d\n", atomic_read(&adapter->pend_cmds)); atomic_read 3811 drivers/scsi/megaraid.c while (atomic_read(&adapter->pend_cmds) > 0 || atomic_read 4113 drivers/scsi/megaraid.c if (atomic_read(&adapter->quiescent) == 0) atomic_read 4533 drivers/scsi/megaraid.c if (atomic_read(&adapter->pend_cmds) > 0) atomic_read 3524 drivers/scsi/megaraid/megaraid_mbox.c if (atomic_read(&adapter->being_detached)) { atomic_read 579 drivers/scsi/megaraid/megaraid_sas_base.c if ((atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) && atomic_read 717 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) atomic_read 856 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) atomic_read 1052 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) atomic_read 1099 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 1128 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 1195 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 1657 drivers/scsi/megaraid/megaraid_sas_base.c dev_err(&instance->pdev->dev, "[%d]: Total OS Pending cmds : %d\n",instance->host->host_no,atomic_read(&instance->fw_outstanding)); atomic_read 1777 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_ADPRESET_SM_INFAULT) { atomic_read 1788 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 1801 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) atomic_read 2188 drivers/scsi/megaraid/megaraid_sas_base.c && atomic_read(&instance->fw_outstanding) < atomic_read 2216 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) atomic_read 2684 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 2690 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) { atomic_read 2701 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HBA_OPERATIONAL) atomic_read 2705 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) { atomic_read 2745 drivers/scsi/megaraid/megaraid_sas_base.c outstanding = atomic_read(&instance->fw_outstanding); atomic_read 2764 drivers/scsi/megaraid/megaraid_sas_base.c outstanding = atomic_read(&instance->fw_outstanding); atomic_read 2773 drivers/scsi/megaraid/megaraid_sas_base.c if ((fw_state == MFI_STATE_FAULT) || atomic_read(&instance->fw_outstanding)) { atomic_read 2776 drivers/scsi/megaraid/megaraid_sas_base.c __func__, __LINE__, fw_state, atomic_read(&instance->fw_outstanding)); atomic_read 2781 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 2792 drivers/scsi/megaraid/megaraid_sas_base.c outstanding = atomic_read(&instance->fw_outstanding); atomic_read 2813 drivers/scsi/megaraid/megaraid_sas_base.c atomic_read(&instance->fw_outstanding)); atomic_read 2838 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 3006 drivers/scsi/megaraid/megaraid_sas_base.c atomic_read(&instance->fw_outstanding)); atomic_read 3298 drivers/scsi/megaraid/megaraid_sas_base.c return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&instance->ldio_outstanding)); atomic_read 3308 drivers/scsi/megaraid/megaraid_sas_base.c return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&instance->fw_outstanding)); atomic_read 3820 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) != MEGASAS_ADPRESET_SM_INFAULT) { atomic_read 3822 drivers/scsi/megaraid/megaraid_sas_base.c atomic_read(&instance->adprecovery)); atomic_read 3826 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_ADPRESET_SM_INFAULT) { atomic_read 3940 drivers/scsi/megaraid/megaraid_sas_base.c fw_state, atomic_read(&instance->adprecovery)); atomic_read 3964 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->fw_reset_no_pci_access)) atomic_read 7442 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) atomic_read 7485 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) atomic_read 7733 drivers/scsi/megaraid/megaraid_sas_base.c adp_state = atomic_read(&instance->adprecovery); atomic_read 7995 drivers/scsi/megaraid/megaraid_sas_base.c if ((atomic_read(&local_instance->adprecovery) == atomic_read 8283 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 8333 drivers/scsi/megaraid/megaraid_sas_base.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 1387 drivers/scsi/megaraid/megaraid_sas_fp.c pend0 = atomic_read(&lbInfo->scsi_pending_cmds[pd0]); atomic_read 1388 drivers/scsi/megaraid/megaraid_sas_fp.c pend1 = atomic_read(&lbInfo->scsi_pending_cmds[pd1]); atomic_read 2830 drivers/scsi/megaraid/megaraid_sas_fusion.c atomic_read(&scp->device->device_busy) > atomic_read 3163 drivers/scsi/megaraid/megaraid_sas_fusion.c atomic_read(&scmd->device->device_busy) > MR_DEVICE_HIGH_IOPS_DEPTH) atomic_read 3534 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) atomic_read 3771 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) atomic_read 4173 drivers/scsi/megaraid/megaraid_sas_fusion.c outstanding = atomic_read(&instance->fw_outstanding); atomic_read 4191 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->fw_outstanding)) { atomic_read 4624 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) { atomic_read 4705 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) { atomic_read 4781 drivers/scsi/megaraid/megaraid_sas_fusion.c (atomic_read(&peer_instance->adprecovery) == atomic_read 4809 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) { atomic_read 4836 drivers/scsi/megaraid/megaraid_sas_fusion.c } while ((atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) && atomic_read 4839 drivers/scsi/megaraid/megaraid_sas_fusion.c if (atomic_read(&instance->adprecovery) == MEGASAS_HBA_OPERATIONAL) { atomic_read 1838 drivers/scsi/mpt3sas/mpt3sas_base.c atomic_read(&ioc->chain_lookup[smid - 1].chain_offset); atomic_read 3488 drivers/scsi/mpt3sas/mpt3sas_base.c if (atomic_read(&scmd->device->device_busy) > atomic_read 2996 drivers/scsi/mpt3sas/mpt3sas_scsih.c if (r == SUCCESS && atomic_read(&scmd->device->device_busy)) atomic_read 3074 drivers/scsi/mpt3sas/mpt3sas_scsih.c if (r == SUCCESS && atomic_read(&starget->target_busy)) atomic_read 416 drivers/scsi/mvumi.c if (atomic_read(&mhba->fw_outstanding) >= mhba->max_io) { atomic_read 420 drivers/scsi/mvumi.c return mhba->max_io - atomic_read(&mhba->fw_outstanding); atomic_read 427 drivers/scsi/mvumi.c if (atomic_read(&mhba->fw_outstanding) >= (mhba->max_io - 1)) atomic_read 717 drivers/scsi/mvumi.c if (atomic_read(&cmd->sync_cmd)) { atomic_read 720 drivers/scsi/mvumi.c atomic_read(&cmd->sync_cmd)); atomic_read 734 drivers/scsi/mvumi.c if (atomic_read(&cmd->sync_cmd)) { atomic_read 1339 drivers/scsi/mvumi.c if (atomic_read(&cmd->sync_cmd)) { atomic_read 1614 drivers/scsi/mvumi.c if (!atomic_read(&mhba->pnp_count)) atomic_read 1761 drivers/scsi/pmcraid.c atomic_read(&pinstance->ccn.ignore) == 1) { atomic_read 1802 drivers/scsi/pmcraid.c atomic_read(&pinstance->ccn.ignore) == 1) { atomic_read 3076 drivers/scsi/pmcraid.c if (atomic_read(&pinstance->outstanding_cmds) <= atomic_read 4302 drivers/scsi/pmcraid.c if (!atomic_read(&pinstance->expose_resources)) atomic_read 5660 drivers/scsi/pmcraid.c if (atomic_read(&pmcraid_adapter_count) >= PMCRAID_MAX_ADAPTERS) { atomic_read 5663 drivers/scsi/pmcraid.c atomic_read(&pmcraid_adapter_count)); atomic_read 5679 drivers/scsi/pmcraid.c atomic_read(&pmcraid_adapter_count)); atomic_read 330 drivers/scsi/qedf/qedf_debugfs.c seq_printf(s, "Link State: %s\n", atomic_read(&qedf->link_state) ? atomic_read 347 drivers/scsi/qedf/qedf_debugfs.c atomic_read(&qedf->cmd_mgr->free_list_cnt)); atomic_read 362 drivers/scsi/qedf/qedf_debugfs.c atomic_read(&fcport->free_sqes), atomic_read 363 drivers/scsi/qedf/qedf_debugfs.c atomic_read(&fcport->num_active_ios)); atomic_read 61 drivers/scsi/qedf/qedf_fip.c if (atomic_read(&qedf->link_state) != QEDF_LINK_UP) { atomic_read 103 drivers/scsi/qedf/qedf_fip.c if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) { atomic_read 296 drivers/scsi/qedf/qedf_io.c atomic_read(&cmgr->free_list_cnt)); atomic_read 316 drivers/scsi/qedf/qedf_io.c free_sqes = atomic_read(&fcport->free_sqes); atomic_read 326 drivers/scsi/qedf/qedf_io.c if ((atomic_read(&fcport->num_active_ios) >= atomic_read 330 drivers/scsi/qedf/qedf_io.c atomic_read(&fcport->num_active_ios)); atomic_read 335 drivers/scsi/qedf/qedf_io.c if (atomic_read(&cmd_mgr->free_list_cnt) <= GBL_RSVD_TASKS) { atomic_read 338 drivers/scsi/qedf/qedf_io.c atomic_read(&cmd_mgr->free_list_cnt)); atomic_read 461 drivers/scsi/qedf/qedf_io.c if (atomic_read(&fcport->num_active_ios) < 0) { atomic_read 1002 drivers/scsi/qedf/qedf_io.c atomic_read(&qedf->link_state) != QEDF_LINK_UP) { atomic_read 1597 drivers/scsi/qedf/qedf_io.c while (atomic_read(&fcport->ios_to_queue)) { atomic_read 1600 drivers/scsi/qedf/qedf_io.c atomic_read(&fcport->ios_to_queue)); atomic_read 1604 drivers/scsi/qedf/qedf_io.c atomic_read(&fcport->ios_to_queue)); atomic_read 1615 drivers/scsi/qedf/qedf_io.c atomic_read(&fcport->num_active_ios), fcport, atomic_read 1666 drivers/scsi/qedf/qedf_io.c if (atomic_read(&io_req->state) == atomic_read 1781 drivers/scsi/qedf/qedf_io.c flush_cnt, atomic_read(&fcport->num_active_ios)); atomic_read 1785 drivers/scsi/qedf/qedf_io.c while (atomic_read(&fcport->num_active_ios)) { atomic_read 1789 drivers/scsi/qedf/qedf_io.c atomic_read(&fcport->num_active_ios), atomic_read 1795 drivers/scsi/qedf/qedf_io.c atomic_read(&fcport->num_active_ios)); atomic_read 1870 drivers/scsi/qedf/qedf_io.c if (atomic_read(&qedf->link_down_tmo_valid) > 0) { atomic_read 1877 drivers/scsi/qedf/qedf_io.c if (!atomic_read(&fcport->free_sqes)) { atomic_read 2172 drivers/scsi/qedf/qedf_io.c if (!atomic_read(&fcport->free_sqes)) { atomic_read 129 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) { atomic_read 143 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_UP) atomic_read 165 drivers/scsi/qedf/qedf_main.c atomic_read(&qedf->link_state)); atomic_read 167 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_UP) { atomic_read 172 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) != QEDF_LINK_UP) { atomic_read 195 drivers/scsi/qedf/qedf_main.c } else if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) { atomic_read 521 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_UP) { atomic_read 535 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->dcbx) == QEDF_DCBX_DONE || atomic_read 539 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_down_tmo_valid) > 0) atomic_read 581 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->dcbx) == QEDF_DCBX_DONE) { atomic_read 607 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_UP && atomic_read 609 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_down_tmo_valid) > 0) atomic_read 813 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->num_offloads)) atomic_read 816 drivers/scsi/qedf/qedf_main.c atomic_read(&qedf->num_offloads)); atomic_read 862 drivers/scsi/qedf/qedf_main.c WARN_ON(atomic_read(&qedf->num_offloads)); atomic_read 894 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN || atomic_read 1060 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) != QEDF_LINK_UP) { atomic_read 1412 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->num_offloads) >= QEDF_MAX_SESSIONS) { atomic_read 1735 drivers/scsi/qedf/qedf_main.c if (atomic_read(&base_qedf->link_state) != QEDF_LINK_UP) { atomic_read 2529 drivers/scsi/qedf/qedf_main.c if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) { atomic_read 3668 drivers/scsi/qedf/qedf_main.c if (!atomic_read(&pdev->enable_cnt)) atomic_read 869 drivers/scsi/qedi/qedi_iscsi.c if (atomic_read(&qedi->link_state) != QEDI_LINK_UP) { atomic_read 25 drivers/scsi/qedi/qedi_sysfs.c if (atomic_read(&qedi->link_state) == QEDI_LINK_UP) atomic_read 1140 drivers/scsi/qla2xxx/qla_attr.c if (atomic_read(&vha->loop_state) == LOOP_DOWN || atomic_read 1141 drivers/scsi/qla2xxx/qla_attr.c atomic_read(&vha->loop_state) == LOOP_DEAD || atomic_read 1144 drivers/scsi/qla2xxx/qla_attr.c else if (atomic_read(&vha->loop_state) != LOOP_READY || atomic_read 2636 drivers/scsi/qla2xxx/qla_attr.c } else if (atomic_read(&base_vha->loop_state) == LOOP_READY && atomic_read 2753 drivers/scsi/qla2xxx/qla_attr.c switch (atomic_read(&base_vha->loop_state)) { atomic_read 2814 drivers/scsi/qla2xxx/qla_attr.c if (atomic_read(&base_vha->loop_state) == LOOP_DOWN || atomic_read 2815 drivers/scsi/qla2xxx/qla_attr.c atomic_read(&base_vha->loop_state) == LOOP_DEAD) { atomic_read 797 drivers/scsi/qla2xxx/qla_bsg.c if (atomic_read(&vha->loop_state) == LOOP_READY && atomic_read 1333 drivers/scsi/qla2xxx/qla_bsg.c if (atomic_read(&fcport->state) != FCS_ONLINE) { atomic_read 4616 drivers/scsi/qla2xxx/qla_def.h atomic_read(&ha->loop_state) == LOOP_DOWN) atomic_read 525 drivers/scsi/qla2xxx/qla_dfs.c if (atomic_read(&qla2x00_dfs_root_count) == 0 && atomic_read 3677 drivers/scsi/qla2xxx/qla_gs.c atomic_read(&fcport->state) == FCS_ONLINE) { atomic_read 4326 drivers/scsi/qla2xxx/qla_gs.c ls = atomic_read(&vha->loop_state); atomic_read 4455 drivers/scsi/qla2xxx/qla_gs.c ls = atomic_read(&vha->loop_state); atomic_read 4237 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&vha->loop_down_timer) && atomic_read 4309 drivers/scsi/qla2xxx/qla_init.c if (LOOP_TRANSITION(vha) || atomic_read(&ha->loop_down_timer) || atomic_read 4807 drivers/scsi/qla2xxx/qla_init.c old_state = atomic_read(&fcport->state); atomic_read 4978 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&vha->loop_down_timer) || atomic_read 5240 drivers/scsi/qla2xxx/qla_init.c atomic_read(&fcport->state) == FCS_ONLINE) { atomic_read 5283 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&fcport->state) != FCS_ONLINE) atomic_read 5331 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&fcport->state) == FCS_ONLINE) atomic_read 5720 drivers/scsi/qla2xxx/qla_init.c (atomic_read(&vha->loop_down_timer) || atomic_read 5830 drivers/scsi/qla2xxx/qla_init.c (atomic_read(&fcport->state) == FCS_ONLINE || atomic_read 5918 drivers/scsi/qla2xxx/qla_init.c atomic_read(&fcport->state) == FCS_ONLINE) { atomic_read 6189 drivers/scsi/qla2xxx/qla_init.c } while (!atomic_read(&vha->loop_down_timer) && atomic_read 6251 drivers/scsi/qla2xxx/qla_init.c atomic_read(&fcport->state) != FCS_UNCONFIGURED) { atomic_read 6598 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 6604 drivers/scsi/qla2xxx/qla_init.c if (!atomic_read(&vha->loop_down_timer)) atomic_read 6659 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&ha->num_pend_mbx_stage3)) { atomic_read 6665 drivers/scsi/qla2xxx/qla_init.c while (atomic_read(&ha->num_pend_mbx_stage3) || atomic_read 6666 drivers/scsi/qla2xxx/qla_init.c atomic_read(&ha->num_pend_mbx_stage2) || atomic_read 6667 drivers/scsi/qla2xxx/qla_init.c atomic_read(&ha->num_pend_mbx_stage1)) { atomic_read 6676 drivers/scsi/qla2xxx/qla_init.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 6692 drivers/scsi/qla2xxx/qla_init.c if (!atomic_read(&vha->loop_down_timer)) atomic_read 6802 drivers/scsi/qla2xxx/qla_init.c if (!atomic_read(&vha->loop_down_timer)) { atomic_read 8583 drivers/scsi/qla2xxx/qla_init.c if (!atomic_read(&vha->loop_down_timer)) { atomic_read 788 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 838 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 881 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 921 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 923 drivers/scsi/qla2xxx/qla_isr.c if (!atomic_read(&vha->loop_down_timer)) atomic_read 951 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 953 drivers/scsi/qla2xxx/qla_isr.c if (!atomic_read(&vha->loop_down_timer)) atomic_read 1019 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&fcport->state) != FCS_ONLINE) atomic_read 1032 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 1058 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN && atomic_read 1060 drivers/scsi/qla2xxx/qla_isr.c atomic_read(&vha->loop_state) != LOOP_DEAD) { atomic_read 1208 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&vha->loop_state) == LOOP_DOWN) atomic_read 2731 drivers/scsi/qla2xxx/qla_isr.c if (atomic_read(&fcport->state) == FCS_ONLINE) { atomic_read 81 drivers/scsi/qla2xxx/qla_mid.c !atomic_read(&vha->vref_count), HZ) > 0) atomic_read 86 drivers/scsi/qla2xxx/qla_mid.c if (atomic_read(&vha->vref_count)) { atomic_read 197 drivers/scsi/qla2xxx/qla_mid.c if (atomic_read(&base_vha->loop_state) == LOOP_DOWN || atomic_read 198 drivers/scsi/qla2xxx/qla_mid.c atomic_read(&base_vha->loop_state) == LOOP_DEAD || atomic_read 328 drivers/scsi/qla2xxx/qla_mid.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 332 drivers/scsi/qla2xxx/qla_mid.c if (!atomic_read(&vha->loop_down_timer)) atomic_read 375 drivers/scsi/qla2xxx/qla_mid.c atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 1133 drivers/scsi/qla2xxx/qla_mr.c if ((atomic_read(&vha->loop_down_timer) || atomic_read 1178 drivers/scsi/qla2xxx/qla_mr.c atomic_read(&fcport->state) == FCS_ONLINE) atomic_read 1189 drivers/scsi/qla2xxx/qla_mr.c atomic_read(&fcport->state), atomic_read 1202 drivers/scsi/qla2xxx/qla_mr.c if (atomic_read(&fcport->state) != FCS_ONLINE) { atomic_read 1275 drivers/scsi/qla2xxx/qla_mr.c if (atomic_read(&fcport->state) == FCS_DEVICE_LOST) { atomic_read 1372 drivers/scsi/qla2xxx/qla_mr.c if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_read 1377 drivers/scsi/qla2xxx/qla_mr.c if (!atomic_read(&vha->loop_down_timer)) atomic_read 1385 drivers/scsi/qla2xxx/qla_mr.c if (atomic_read(&fcport->state) == FCS_ONLINE) atomic_read 2513 drivers/scsi/qla2xxx/qla_mr.c atomic_read(&fcport->state)); atomic_read 2515 drivers/scsi/qla2xxx/qla_mr.c if (atomic_read(&fcport->state) == FCS_ONLINE) atomic_read 867 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) != FCS_ONLINE) { atomic_read 868 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD || atomic_read 869 drivers/scsi/qla2xxx/qla_os.c atomic_read(&base_vha->loop_state) == LOOP_DEAD) { atomic_read 872 drivers/scsi/qla2xxx/qla_os.c atomic_read(&fcport->state), atomic_read 873 drivers/scsi/qla2xxx/qla_os.c atomic_read(&base_vha->loop_state)); atomic_read 949 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) != FCS_ONLINE) { atomic_read 950 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD || atomic_read 951 drivers/scsi/qla2xxx/qla_os.c atomic_read(&base_vha->loop_state) == LOOP_DEAD) { atomic_read 954 drivers/scsi/qla2xxx/qla_os.c atomic_read(&fcport->state), atomic_read 955 drivers/scsi/qla2xxx/qla_os.c atomic_read(&base_vha->loop_state)); atomic_read 2723 drivers/scsi/qla2xxx/qla_os.c return atomic_read(&vha->loop_state) == LOOP_READY; atomic_read 3524 drivers/scsi/qla2xxx/qla_os.c if (!atomic_read(&pdev->enable_cnt)) atomic_read 3692 drivers/scsi/qla2xxx/qla_os.c if (!atomic_read(&pdev->enable_cnt)) { atomic_read 3886 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) == FCS_ONLINE && atomic_read 3895 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) != FCS_DEVICE_DEAD) atomic_read 3936 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD) atomic_read 3938 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) == FCS_ONLINE) { atomic_read 5359 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&fcport->state) != FCS_ONLINE && atomic_read 6060 drivers/scsi/qla2xxx/qla_os.c if (!atomic_read(&pdev->enable_cnt)) { atomic_read 6368 drivers/scsi/qla2xxx/qla_os.c atomic_read(&base_vha->loop_state) != LOOP_DOWN) { atomic_read 6404 drivers/scsi/qla2xxx/qla_os.c atomic_read(&base_vha->loop_state) == LOOP_READY) { atomic_read 6506 drivers/scsi/qla2xxx/qla_os.c !atomic_read(&vha->loop_down_timer) && atomic_read 6516 drivers/scsi/qla2xxx/qla_os.c } while (!atomic_read(&vha->loop_down_timer) && atomic_read 6572 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&vha->loop_down_timer) > 0 && atomic_read 6577 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&vha->loop_down_timer) == atomic_read 6641 drivers/scsi/qla2xxx/qla_os.c atomic_read(&vha->loop_down_timer)); atomic_read 6670 drivers/scsi/qla2xxx/qla_os.c (atomic_read(&ha->nvme_active_aen_cnt) != ha->nvme_last_rptd_aen) && atomic_read 6676 drivers/scsi/qla2xxx/qla_os.c ha->nvme_last_rptd_aen = atomic_read(&ha->nvme_active_aen_cnt); atomic_read 6682 drivers/scsi/qla2xxx/qla_os.c (atomic_read(&ha->zio_threshold) != ha->last_zio_threshold) && atomic_read 6688 drivers/scsi/qla2xxx/qla_os.c ha->last_zio_threshold = atomic_read(&ha->zio_threshold); atomic_read 6859 drivers/scsi/qla2xxx/qla_os.c if (atomic_read(&ha->num_pend_mbx_stage3)) { atomic_read 6866 drivers/scsi/qla2xxx/qla_os.c while (atomic_read(&ha->num_pend_mbx_stage3) || atomic_read 6867 drivers/scsi/qla2xxx/qla_os.c atomic_read(&ha->num_pend_mbx_stage2) || atomic_read 6868 drivers/scsi/qla2xxx/qla_os.c atomic_read(&ha->num_pend_mbx_stage1)) { atomic_read 6920 drivers/scsi/qla2xxx/qla_os.c if (!atomic_read(&pdev->enable_cnt)) { atomic_read 6114 drivers/scsi/qla2xxx/qla_target.c atomic_read(&vha->vha_tgt.qla_tgt->tgt_global_resets_count); atomic_read 6143 drivers/scsi/qla2xxx/qla_target.c atomic_read(&vha->vha_tgt.qla_tgt->tgt_global_resets_count)) { atomic_read 6148 drivers/scsi/qla2xxx/qla_target.c atomic_read(&vha->vha_tgt. atomic_read 892 drivers/scsi/qla2xxx/tcm_qla2xxx.c atomic_read(&tpg->lport_tpg_enabled)); atomic_read 918 drivers/scsi/qla2xxx/tcm_qla2xxx.c if (atomic_read(&tpg->lport_tpg_enabled)) atomic_read 924 drivers/scsi/qla2xxx/tcm_qla2xxx.c if (!atomic_read(&tpg->lport_tpg_enabled)) atomic_read 1080 drivers/scsi/qla2xxx/tcm_qla2xxx.c if (atomic_read(&tpg->lport_tpg_enabled)) atomic_read 1086 drivers/scsi/qla2xxx/tcm_qla2xxx.c if (!atomic_read(&tpg->lport_tpg_enabled)) atomic_read 1691 drivers/scsi/qla2xxx/tcm_qla2xxx.c !atomic_read(&base_lport->tpg_1->lport_tpg_enabled)) { atomic_read 4477 drivers/scsi/qla4xxx/ql4_os.c if (atomic_read(&ddb_entry->retry_relogin_timer) != atomic_read 4479 drivers/scsi/qla4xxx/ql4_os.c if (atomic_read(&ddb_entry->retry_relogin_timer) == atomic_read 4494 drivers/scsi/qla4xxx/ql4_os.c if (atomic_read(&ddb_entry->relogin_timer) && atomic_read 4507 drivers/scsi/qla4xxx/ql4_os.c atomic_read(&ddb_entry->relogin_retry_count), atomic_read 8893 drivers/scsi/qla4xxx/ql4_os.c if (atomic_read(&other_pdev->enable_cnt)) { atomic_read 9664 drivers/scsi/qla4xxx/ql4_os.c if (atomic_read(&other_pdev->enable_cnt)) { atomic_read 195 drivers/scsi/scsi.c if (atomic_read(&shost->host_blocked)) atomic_read 197 drivers/scsi/scsi.c if (atomic_read(&starget->target_blocked)) atomic_read 199 drivers/scsi/scsi.c if (atomic_read(&sdev->device_blocked)) atomic_read 1617 drivers/scsi/scsi_debug.c changing = atomic_read(&devip->stopped) == !stop; atomic_read 3747 drivers/scsi/scsi_debug.c if (unlikely(atomic_read(&retired_max_queue) > 0)) atomic_read 3760 drivers/scsi/scsi_debug.c retval = atomic_read(&retired_max_queue); atomic_read 3935 drivers/scsi/scsi_debug.c r_qmax = atomic_read(&retired_max_queue); atomic_read 4216 drivers/scsi/scsi_debug.c count = atomic_read(&sdebug_cmnd_count); atomic_read 4232 drivers/scsi/scsi_debug.c if ((atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth)) > 0) { atomic_read 4280 drivers/scsi/scsi_debug.c if (unlikely(atomic_read(&sqp->blocked))) { atomic_read 4284 drivers/scsi/scsi_debug.c num_in_q = atomic_read(&devip->num_in_q); atomic_read 4598 drivers/scsi/scsi_debug.c atomic_read(&sdebug_cmnd_count), atomic_read 4599 drivers/scsi/scsi_debug.c atomic_read(&sdebug_completions), atomic_read 4600 drivers/scsi/scsi_debug.c "miss_cpus", atomic_read(&sdebug_miss_cpus), atomic_read 4601 drivers/scsi/scsi_debug.c atomic_read(&sdebug_a_tsf)); atomic_read 5552 drivers/scsi/scsi_debug.c num_in_q = atomic_read(&devip->num_in_q); atomic_read 5571 drivers/scsi/scsi_debug.c if (0 == (atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth))) { atomic_read 5586 drivers/scsi/scsi_debug.c (atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth)) == 0; atomic_read 5705 drivers/scsi/scsi_debug.c if (unlikely((F_M_ACCESS & flags) && atomic_read(&devip->stopped))) { atomic_read 413 drivers/scsi/scsi_lib.c if (atomic_read(&sdev->device_busy) >= sdev->queue_depth) atomic_read 415 drivers/scsi/scsi_lib.c if (atomic_read(&sdev->device_blocked) > 0) atomic_read 423 drivers/scsi/scsi_lib.c if (atomic_read(&starget->target_busy) >= starget->can_queue) atomic_read 425 drivers/scsi/scsi_lib.c if (atomic_read(&starget->target_blocked) > 0) atomic_read 434 drivers/scsi/scsi_lib.c atomic_read(&shost->host_busy) >= shost->can_queue) atomic_read 436 drivers/scsi/scsi_lib.c if (atomic_read(&shost->host_blocked) > 0) atomic_read 1287 drivers/scsi/scsi_lib.c if (atomic_read(&sdev->device_blocked)) { atomic_read 1334 drivers/scsi/scsi_lib.c if (atomic_read(&starget->target_blocked) > 0) { atomic_read 1378 drivers/scsi/scsi_lib.c if (atomic_read(&shost->host_blocked) > 0) { atomic_read 1639 drivers/scsi/scsi_lib.c if (atomic_read(&sdev->device_busy) == 0 && !scsi_device_blocked(sdev)) atomic_read 1710 drivers/scsi/scsi_lib.c if (atomic_read(&sdev->device_busy) || atomic_read 2915 drivers/scsi/scsi_lib.c if (WARN_ON_ONCE(atomic_read(&sdev->disk_events_disable_depth) <= 0)) atomic_read 653 drivers/scsi/scsi_sysfs.c return snprintf(buf, 20, "%d\n", atomic_read(&sdev->device_busy)); atomic_read 662 drivers/scsi/scsi_sysfs.c return snprintf(buf, 20, "%d\n", atomic_read(&sdev->device_blocked)); atomic_read 910 drivers/scsi/scsi_sysfs.c unsigned long long count = atomic_read(&sdev->field); \ atomic_read 1775 drivers/scsi/scsi_transport_iscsi.c return !atomic_read(&ihost->nr_scans); atomic_read 254 drivers/scsi/sg.c (atomic_read(&sdp->detaching) || atomic_read 260 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 267 drivers/scsi/sg.c (atomic_read(&sdp->detaching) || atomic_read 273 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 464 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) { atomic_read 473 drivers/scsi/sg.c (atomic_read(&sdp->detaching) || atomic_read 475 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) { atomic_read 624 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 821 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) { atomic_read 920 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 931 drivers/scsi/sg.c (srp_done(sfp, srp) || atomic_read(&sdp->detaching))); atomic_read 932 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 975 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 1099 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 1103 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 1131 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 1196 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 1339 drivers/scsi/sg.c if (unlikely(atomic_read(&sdp->detaching))) atomic_read 2169 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) { atomic_read 2285 drivers/scsi/sg.c else if (atomic_read(&sdp->detaching)) { atomic_read 2481 drivers/scsi/sg.c (atomic_read(&sdp->detaching))) atomic_read 2490 drivers/scsi/sg.c (int) atomic_read(&scsidp->device_busy), atomic_read 2507 drivers/scsi/sg.c if (sdp && scsidp && (!atomic_read(&sdp->detaching))) atomic_read 2597 drivers/scsi/sg.c if (atomic_read(&sdp->detaching)) atomic_read 288 drivers/scsi/smartpqi/smartpqi_init.c while (atomic_read(&ctrl_info->num_busy_threads) > atomic_read 289 drivers/scsi/smartpqi/smartpqi_init.c atomic_read(&ctrl_info->num_blocked_threads)) atomic_read 1515 drivers/scsi/smartpqi/smartpqi_init.c atomic_read(&device->scsi_cmds_outstanding)); atomic_read 2932 drivers/scsi/smartpqi/smartpqi_init.c WARN_ON(atomic_read(&io_request->refcount) == 0); atomic_read 3218 drivers/scsi/smartpqi/smartpqi_init.c num_interrupts = atomic_read(&ctrl_info->num_interrupts); atomic_read 3244 drivers/scsi/smartpqi/smartpqi_init.c atomic_read(&ctrl_info->num_interrupts); atomic_read 5535 drivers/scsi/smartpqi/smartpqi_init.c while (atomic_read(&device->scsi_cmds_outstanding)) { atomic_read 5567 drivers/scsi/smartpqi/smartpqi_init.c if (atomic_read(&device->scsi_cmds_outstanding)) { atomic_read 5597 drivers/scsi/smartpqi/smartpqi_init.c while (atomic_read(&ctrl_info->sync_cmds_outstanding)) { atomic_read 7793 drivers/scsi/smartpqi/smartpqi_init.c if (atomic_read(&io_request->refcount) == 0) atomic_read 7934 drivers/scsi/smartpqi/smartpqi_init.c if (atomic_read(&io_request->refcount) == 0) atomic_read 343 drivers/scsi/snic/snic_main.c return atomic_read(&snic->state); atomic_read 2355 drivers/scsi/snic/snic_scsi.c while (atomic_read(&snic->ios_inflight)) atomic_read 611 drivers/scsi/sr.c if (!atomic_read(&cd->device->disk_events_disable_depth)) atomic_read 486 drivers/scsi/st.c atomic64_add(atomic_read(&STp->stats->last_write_size) atomic_read 492 drivers/scsi/st.c atomic64_add(atomic_read(&STp->stats->last_write_size), atomic_read 500 drivers/scsi/st.c atomic64_add(atomic_read(&STp->stats->last_read_size) atomic_read 506 drivers/scsi/st.c atomic64_add(atomic_read(&STp->stats->last_read_size), atomic_read 596 drivers/scsi/storvsc_drv.c atomic_read(&dev->num_outstanding_req) == 0); atomic_read 616 drivers/scsi/storvsc_drv.c (atomic_read(&stor_device->num_outstanding_req) == 0)) atomic_read 423 drivers/sh/maple/maple.c if (mdev->interval > 0 && atomic_read(&mdev->busy) == 0 && atomic_read 436 drivers/sh/maple/maple.c if (atomic_read(&mdev->busy) == 0) { atomic_read 125 drivers/soc/fsl/qbman/qman_test_stash.c while (!atomic_read(&bstrap.started)) atomic_read 353 drivers/soc/ti/knav_dma.c if (atomic_read(&chan->ref_count)) atomic_read 363 drivers/soc/ti/knav_dma.c if (atomic_read(&dma->ref_count)) { atomic_read 489 drivers/soc/ti/knav_dma.c if (atomic_read(&chan->ref_count) >= 1) { atomic_read 64 drivers/soc/ti/knav_qmss_acc.c if (!enabled || atomic_read(&kq->desc_count) <= 0) atomic_read 110 drivers/soc/ti/knav_qmss_acc.c if (atomic_read(&acc->retrigger_count)) { atomic_read 90 drivers/soc/ti/knav_qmss_queue.c if (atomic_read(&qh->notifier_enabled) <= 0) atomic_read 422 drivers/soc/ti/knav_qmss_queue.c atomic_read(&inst->desc_count); atomic_read 562 drivers/soc/ti/knav_qmss_queue.c while (atomic_read(&qh->notifier_enabled) > 0) atomic_read 1176 drivers/spi/spi-pxa2xx.c if (atomic_read(&drv_data->dma_running)) atomic_read 260 drivers/spi/spi-rockchip.c if (atomic_read(&rs->state) & TXDMA) atomic_read 263 drivers/spi/spi-rockchip.c if (atomic_read(&rs->state) & RXDMA) atomic_read 763 drivers/staging/android/ashmem.c wait_event(ashmem_shrink_wait, !atomic_read(&ashmem_shrink_inflight)); atomic_read 456 drivers/staging/android/vsoc.c if (atomic_read(address) != arg->value) { atomic_read 596 drivers/staging/android/vsoc.c (atomic_read(reg_data->incoming_signalled) != 0)); atomic_read 556 drivers/staging/fieldbus/anybuss/host.c if (atomic_read(&cd->ind_ab) & pd->flags) { atomic_read 574 drivers/staging/fieldbus/anybuss/host.c if (!(atomic_read(&cd->ind_ab) & pd->flags)) { atomic_read 704 drivers/staging/fieldbus/anybuss/host.c if (((atomic_read(&cd->ind_ab) ^ ind_ap) & IND_AX_MOUT) == 0) { atomic_read 729 drivers/staging/fieldbus/anybuss/host.c if ((atomic_read(&cd->ind_ab) ^ ind_ap) & IND_AX_MIN) { atomic_read 908 drivers/staging/fieldbus/anybuss/host.c ind_ap |= atomic_read(&cd->ind_ab) & IND_AX_EVNT; atomic_read 925 drivers/staging/fieldbus/anybuss/host.c if (!((atomic_read(&cd->ind_ab) ^ ind_ap) & IND_AX_EVNT)) atomic_read 988 drivers/staging/fieldbus/anybuss/host.c ind_ab = atomic_read(&cd->ind_ab); atomic_read 992 drivers/staging/fieldbus/anybuss/host.c (atomic_read(&cd->ind_ab) != ind_ab) || atomic_read 420 drivers/staging/greybus/loopback.c !atomic_read(&gb->outstanding_operations)); atomic_read 829 drivers/staging/greybus/loopback.c (atomic_read(&gb->outstanding_operations) < atomic_read 775 drivers/staging/isdn/gigaset/bas-gigaset.c atomic_read(&bcs->hw.bas->corrbytes)); atomic_read 1098 drivers/staging/isdn/gigaset/bas-gigaset.c corrbytes = atomic_read(&ubc->corrbytes); atomic_read 168 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->sleepstatus.status) == 0) { atomic_read 179 drivers/staging/ks7010/ks7010_sdio.c priv->sleep_mode = atomic_read(&priv->sleepstatus.status); atomic_read 189 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->sleepstatus.status) == 1) { atomic_read 201 drivers/staging/ks7010/ks7010_sdio.c priv->sleep_mode = atomic_read(&priv->sleepstatus.status); atomic_read 208 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) { atomic_read 235 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) atomic_read 244 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.status), atomic_read 245 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.confirm_wait), atomic_read 246 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.snooze_guard), atomic_read 249 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.confirm_wait) || atomic_read 250 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.snooze_guard) || atomic_read 356 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.status) == PS_SNOOZE) atomic_read 466 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.confirm_wait) && is_hif_conf(event)) { atomic_read 501 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) { atomic_read 510 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->sleepstatus.doze_request) == 1) { atomic_read 515 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->sleepstatus.wakeup_request) == 1) { atomic_read 524 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.status)); atomic_read 565 drivers/staging/ks7010/ks7010_sdio.c atomic_read(&priv->psstatus.status) == PS_SNOOZE) { atomic_read 572 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) { atomic_read 592 drivers/staging/ks7010/ks7010_sdio.c if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) { atomic_read 74 drivers/staging/ks7010/ks_hostif.c ps_status = atomic_read(&priv->psstatus.status); atomic_read 89 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->sme_task.count) > 0) atomic_read 666 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->psstatus.snooze_guard)) atomic_read 670 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->psstatus.snooze_guard)) atomic_read 1098 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) { atomic_read 1239 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->psstatus.status) > PS_ACTIVE_SET) atomic_read 1591 drivers/staging/ks7010/ks_hostif.c priv->dev_data[atomic_read(&priv->rec_count)] = p; atomic_read 1592 drivers/staging/ks7010/ks_hostif.c priv->dev_size[atomic_read(&priv->rec_count)] = size; atomic_read 1594 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->event_count) != DEVICE_STOCK_COUNT) { atomic_read 1599 drivers/staging/ks7010/ks_hostif.c if (atomic_read(&priv->rec_count) == DEVICE_STOCK_COUNT) atomic_read 69 drivers/staging/ks7010/ks_wlan_net.c if (atomic_read(&update_phyinfo)) atomic_read 640 drivers/staging/ks7010/ks_wlan_net.c in_interrupt(), atomic_read(&update_phyinfo)); atomic_read 646 drivers/staging/ks7010/ks_wlan_net.c if (!atomic_read(&update_phyinfo)) atomic_read 1746 drivers/staging/ks7010/ks_wlan_net.c if (!atomic_read(&update_phyinfo)) atomic_read 99 drivers/staging/media/ipu3/ipu3-v4l2.c if (atomic_read(&imgu_sd->running_mode) == IPU3_RUNNING_MODE_VIDEO) atomic_read 671 drivers/staging/media/ipu3/ipu3-v4l2.c if (atomic_read(&imgu_sd->running_mode) == IPU3_RUNNING_MODE_VIDEO) atomic_read 572 drivers/staging/media/ipu3/ipu3.c if (!atomic_read(&imgu->qbuf_barrier)) atomic_read 196 drivers/staging/media/meson/vdec/esparser.c atomic_read(&sess->esparser_queued_bufs) >= num_dst_bufs) atomic_read 276 drivers/staging/media/meson/vdec/vdec_helpers.c atomic_read(&sess->esparser_queued_bufs) <= 2) { atomic_read 284 drivers/staging/media/meson/vdec/vdec_helpers.c atomic_read(&sess->esparser_queued_bufs)); atomic_read 657 drivers/staging/media/omap4iss/iss.c if (!wait_event_timeout(*wait, !atomic_read(stopping), atomic_read 459 drivers/staging/media/omap4iss/iss_video.c buf->vb.sequence = atomic_read(&pipe->frame_number); atomic_read 435 drivers/staging/octeon/ethernet-rx.c if (!atomic_read(&oct_rx_ready)) atomic_read 120 drivers/staging/octeon/ethernet.c if (!atomic_read(&cvm_oct_poll_queue_stopping)) atomic_read 136 drivers/staging/octeon/ethernet.c if (!atomic_read(&cvm_oct_poll_queue_stopping)) atomic_read 1672 drivers/staging/qlge/qlge_dbg.c pr_err("tx_ring->tx_count = %d\n", atomic_read(&tx_ring->tx_count)); atomic_read 554 drivers/staging/qlge/qlge_ethtool.c return atomic_read(&qdev->lb_count) ? -EIO : 0; atomic_read 673 drivers/staging/qlge/qlge_main.c if (!atomic_read(&ctx->irq_cnt)) { atomic_read 2239 drivers/staging/qlge/qlge_main.c if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4))) atomic_read 2504 drivers/staging/qlge/qlge_main.c if (atomic_read(&qdev->intr_context[0].irq_cnt)) { atomic_read 2648 drivers/staging/qlge/qlge_main.c if (unlikely(atomic_read(&tx_ring->tx_count) < 2)) { atomic_read 2704 drivers/staging/qlge/qlge_main.c if (unlikely(atomic_read(&tx_ring->tx_count) < 2)) { atomic_read 2706 drivers/staging/qlge/qlge_main.c if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4))) atomic_read 721 drivers/staging/ralink-gdma/ralink-gdma.c if (atomic_read(&dma_dev->cnt) >= 2) { atomic_read 44 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c if (atomic_read(&entry->refcnt) != 0 && !force) atomic_read 583 drivers/staging/rtl8192u/r8192U_core.c atomic_read(&(priv->tx_pending[VI_PRIORITY])), atomic_read 584 drivers/staging/rtl8192u/r8192U_core.c atomic_read(&(priv->tx_pending[VO_PRIORITY])), atomic_read 585 drivers/staging/rtl8192u/r8192U_core.c atomic_read(&(priv->tx_pending[BE_PRIORITY])), atomic_read 586 drivers/staging/rtl8192u/r8192U_core.c atomic_read(&(priv->tx_pending[BK_PRIORITY])), atomic_read 652 drivers/staging/rtl8192u/r8192U_core.c int used = atomic_read(&priv->tx_pending[queue_index]); atomic_read 1429 drivers/staging/rtl8192u/r8192U_core.c pend = atomic_read(&priv->tx_pending[tcb_desc->queue_index]); atomic_read 1580 drivers/staging/rtl8192u/r8192U_core.c atomic_read(&priv->tx_pending[tcb_desc->queue_index]), atomic_read 1591 drivers/staging/rtl8192u/r8192U_core.c atomic_read(&priv->tx_pending[tcb_desc->queue_index]), atomic_read 1594 drivers/staging/rtl8192u/r819xU_phy.c atomic_read(&priv->ieee80211->atm_swbw)); atomic_read 320 drivers/staging/rtl8723bs/core/rtw_cmd.c || atomic_read(&(pcmdpriv->cmdthd_running)) == false /* com_thread not running */ atomic_read 393 drivers/staging/rtl8723bs/core/rtw_cmd.c atomic_read(&(adapter->cmdpriv.cmdthd_running)) == true && atomic_read 1923 drivers/staging/rtl8723bs/core/rtw_mlme.c return (atomic_read(&mlmepriv->set_scan_deny) != 0) ? true : false; atomic_read 6618 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c if (evt_seq != (atomic_read(&pevt_priv->event_seq) & 0x7f)) { atomic_read 6621 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c (atomic_read(&pevt_priv->event_seq) & 0x7f))); atomic_read 639 drivers/staging/rtl8723bs/include/drv_types.h int df = atomic_read(&adapter_to_dvobj(padapter)->disable_func); atomic_read 646 drivers/staging/rtl8723bs/include/drv_types.h int df = atomic_read(&adapter_to_dvobj(padapter)->disable_func); atomic_read 651 drivers/staging/rtl8723bs/include/drv_types.h #define RTW_IS_FUNC_DISABLED(padapter, func_bit) (atomic_read(&adapter_to_dvobj(padapter)->disable_func) & (func_bit)) atomic_read 1173 drivers/staging/rtl8723bs/os_dep/os_intfs.c while (atomic_read(&(pcmdpriv->cmdthd_running)) == true) { atomic_read 377 drivers/staging/unisys/visorhba/visorhba_main.c if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT) atomic_read 404 drivers/staging/unisys/visorhba/visorhba_main.c if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT) atomic_read 432 drivers/staging/unisys/visorhba/visorhba_main.c if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT) atomic_read 806 drivers/staging/unisys/visorhba/visorhba_main.c if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT) { atomic_read 884 drivers/staging/unisys/visorhba/visorhba_main.c if (atomic_read(&vdisk->ios_threshold) > 0) { atomic_read 886 drivers/staging/unisys/visorhba/visorhba_main.c if (atomic_read(&vdisk->ios_threshold) == 0) atomic_read 978 drivers/staging/unisys/visorhba/visorhba_main.c devdata->rsp_queue, (atomic_read( atomic_read 559 drivers/staging/unisys/visornic/visornic_main.c if (atomic_read(&devdata->usage) > 1) { atomic_read 565 drivers/staging/unisys/visornic/visornic_main.c if (atomic_read(&devdata->usage)) atomic_read 1514 drivers/staging/unisys/visornic/visornic_main.c atomic_read(&devdata->num_rcvbuf_in_iovm)); atomic_read 769 drivers/staging/uwb/uwb.h unsigned samples = atomic_read(&stats->samples); atomic_read 799 drivers/staging/uwb/uwb.h int samples = atomic_read(&stats->samples); atomic_read 71 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c pos = atomic_read(&alsa_stream->pos); atomic_read 312 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c atomic_read(&alsa_stream->pos)); atomic_read 625 drivers/staging/vc04_services/bcm2835-camera/bcm2835-camera.c while (atomic_read(&port->buffers_with_vpu)) { atomic_read 628 drivers/staging/vc04_services/bcm2835-camera/bcm2835-camera.c __func__, atomic_read(&port->buffers_with_vpu)); atomic_read 634 drivers/staging/vc04_services/bcm2835-camera/bcm2835-camera.c atomic_read(&port->buffers_with_vpu)); atomic_read 513 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c value = atomic_read(&service->poll_flags); atomic_read 518 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c value = atomic_read(&state->poll_services[ atomic_read 176 drivers/staging/wusbcore/wa-nep.c if (atomic_read(&wa->notifs_queued) > 200) { atomic_read 1462 drivers/staging/wusbcore/wa-xfer.c while (atomic_read(&rpipe->segs_available) > 0 atomic_read 1481 drivers/staging/wusbcore/wa-xfer.c atomic_read(&rpipe->segs_available), result); atomic_read 1506 drivers/staging/wusbcore/wa-xfer.c && (atomic_read(&rpipe->segs_available) == atomic_read 1558 drivers/staging/wusbcore/wa-xfer.c BUG_ON(atomic_read(&rpipe->segs_available) > maxrequests); atomic_read 1564 drivers/staging/wusbcore/wa-xfer.c available = atomic_read(&rpipe->segs_available); atomic_read 1604 drivers/staging/wusbcore/wa-xfer.c && (atomic_read(&rpipe->segs_available) == atomic_read 2746 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 2804 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 2959 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3002 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3083 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3223 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3303 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3511 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3552 drivers/target/iscsi/iscsi_target.c hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 3685 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->check_immediate_queue)) atomic_read 3762 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->check_immediate_queue)) atomic_read 4162 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->connection_recovery)) { atomic_read 4176 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->conn_logout_remove)) { atomic_read 4197 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->connection_recovery)) atomic_read 4208 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->sleep_on_conn_wait_comp)) { atomic_read 4221 drivers/target/iscsi/iscsi_target.c if (atomic_read(&conn->connection_wait_rcfr)) { atomic_read 4258 drivers/target/iscsi/iscsi_target.c " %s\n", atomic_read(&sess->nconn), atomic_read 4265 drivers/target/iscsi/iscsi_target.c !atomic_read(&sess->session_logout)) atomic_read 4274 drivers/target/iscsi/iscsi_target.c if (atomic_read(&sess->nconn)) { atomic_read 4275 drivers/target/iscsi/iscsi_target.c if (!atomic_read(&sess->session_reinstatement) && atomic_read 4276 drivers/target/iscsi/iscsi_target.c !atomic_read(&sess->session_fall_back_to_erl0)) { atomic_read 4280 drivers/target/iscsi/iscsi_target.c if (!atomic_read(&sess->session_stop_active)) { atomic_read 4303 drivers/target/iscsi/iscsi_target.c if (!atomic_read(&sess->session_reinstatement) && atomic_read 4304 drivers/target/iscsi/iscsi_target.c atomic_read(&sess->session_fall_back_to_erl0)) { atomic_read 4310 drivers/target/iscsi/iscsi_target.c } else if (atomic_read(&sess->session_logout)) { atomic_read 4314 drivers/target/iscsi/iscsi_target.c if (atomic_read(&sess->session_close)) { atomic_read 4327 drivers/target/iscsi/iscsi_target.c if (!atomic_read(&sess->session_continuation)) atomic_read 4330 drivers/target/iscsi/iscsi_target.c if (atomic_read(&sess->session_close)) { atomic_read 4351 drivers/target/iscsi/iscsi_target.c if (atomic_read(&sess->nconn)) { atomic_read 4353 drivers/target/iscsi/iscsi_target.c " to %s\n", atomic_read(&sess->nconn), atomic_read 4584 drivers/target/iscsi/iscsi_target.c u16 conn_count = atomic_read(&sess->nconn); atomic_read 4618 drivers/target/iscsi/iscsi_target.c if (session_sleep && atomic_read(&sess->nconn)) { atomic_read 4644 drivers/target/iscsi/iscsi_target.c if (atomic_read(&sess->session_fall_back_to_erl0) || atomic_read 4645 drivers/target/iscsi/iscsi_target.c atomic_read(&sess->session_logout) || atomic_read 4646 drivers/target/iscsi/iscsi_target.c atomic_read(&sess->session_close) || atomic_read 554 drivers/target/iscsi/iscsi_target_configfs.c max_cmd_sn = (u32) atomic_read(&sess->max_cmd_sn); atomic_read 1477 drivers/target/iscsi/iscsi_target_configfs.c if (atomic_read(&sess->session_fall_back_to_erl0) || atomic_read 1478 drivers/target/iscsi/iscsi_target_configfs.c atomic_read(&sess->session_logout) || atomic_read 1479 drivers/target/iscsi/iscsi_target_configfs.c atomic_read(&sess->session_close) || atomic_read 755 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&sess->session_reinstatement)) { atomic_read 825 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&conn->connection_exit)) { atomic_read 830 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&conn->transport_failed)) { atomic_read 849 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&conn->connection_exit)) { atomic_read 854 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&conn->transport_failed)) { atomic_read 859 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&conn->connection_reinstatement)) { atomic_read 896 drivers/target/iscsi/iscsi_target_erl0.c !atomic_read(&sess->session_reinstatement) && atomic_read 897 drivers/target/iscsi/iscsi_target_erl0.c !atomic_read(&sess->session_fall_back_to_erl0)) atomic_read 912 drivers/target/iscsi/iscsi_target_erl0.c if (atomic_read(&conn->connection_exit)) { atomic_read 157 drivers/target/iscsi/iscsi_target_login.c if (atomic_read(&sess_p->session_fall_back_to_erl0) || atomic_read 158 drivers/target/iscsi/iscsi_target_login.c atomic_read(&sess_p->session_logout) || atomic_read 159 drivers/target/iscsi/iscsi_target_login.c atomic_read(&sess_p->session_close) || atomic_read 487 drivers/target/iscsi/iscsi_target_login.c if (atomic_read(&sess_p->session_fall_back_to_erl0) || atomic_read 488 drivers/target/iscsi/iscsi_target_login.c atomic_read(&sess_p->session_logout) || atomic_read 489 drivers/target/iscsi/iscsi_target_login.c atomic_read(&sess_p->session_close) || atomic_read 602 drivers/target/iscsi/iscsi_target_login.c if ((atomic_read(&sess->nconn) + 1) > sess->sess_ops->MaxConnections) { atomic_read 718 drivers/target/iscsi/iscsi_target_login.c " from node: %s\n", atomic_read(&sess->nconn), atomic_read 766 drivers/target/iscsi/iscsi_target_login.c " %s\n", atomic_read(&sess->nconn), atomic_read 340 drivers/target/iscsi/iscsi_target_nego.c login_rsp->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn)); atomic_read 45 drivers/target/iscsi/iscsi_target_tmr.c iscsi_sna_lte(be32_to_cpu(hdr->refcmdsn), (u32) atomic_read(&conn->sess->max_cmd_sn))) ? atomic_read 269 drivers/target/iscsi/iscsi_target_util.c max_cmdsn = atomic_read(&sess->max_cmd_sn); atomic_read 571 drivers/target/iscsi/iscsi_target_util.c if (!atomic_read(&cmd->immed_queue_count)) { atomic_read 586 drivers/target/iscsi/iscsi_target_util.c if (atomic_read(&cmd->immed_queue_count)) { atomic_read 589 drivers/target/iscsi/iscsi_target_util.c atomic_read(&cmd->immed_queue_count)); atomic_read 647 drivers/target/iscsi/iscsi_target_util.c if (!atomic_read(&cmd->response_queue_count)) { atomic_read 663 drivers/target/iscsi/iscsi_target_util.c if (atomic_read(&cmd->response_queue_count)) { atomic_read 666 drivers/target/iscsi/iscsi_target_util.c atomic_read(&cmd->response_queue_count)); atomic_read 771 drivers/target/loopback/tcm_loop.c if (atomic_read(&tpg->tl_tpg_port_count)) { atomic_read 773 drivers/target/loopback/tcm_loop.c atomic_read(&tpg->tl_tpg_port_count)); atomic_read 687 drivers/target/target_core_alua.c if (atomic_read(&lun->lun_tg_pt_secondary_offline)) { atomic_read 1213 drivers/target/target_core_alua.c atomic_read(&lun->lun_tg_pt_secondary_offline), atomic_read 1506 drivers/target/target_core_alua.c while (atomic_read(&lu_gp->lu_gp_ref_cnt)) atomic_read 1554 drivers/target/target_core_alua.c while (atomic_read(&lu_gp_mem->lu_gp_mem_ref_cnt)) atomic_read 1762 drivers/target/target_core_alua.c while (atomic_read(&tg_pt_gp->tg_pt_gp_ref_cnt)) atomic_read 1899 drivers/target/target_core_alua.c atomic_read(&lun->lun_tg_pt_secondary_offline) ? atomic_read 2164 drivers/target/target_core_alua.c atomic_read(&lun->lun_tg_pt_secondary_offline)); atomic_read 484 drivers/target/target_core_configfs.c BUG_ON(atomic_read(&t->tf_access_cnt)); atomic_read 278 drivers/target/target_core_iblock.c if (atomic_read(&ibr->ib_bio_err_cnt)) atomic_read 1321 drivers/target/target_core_pr.c while (atomic_read(&pr_reg->pr_res_holders) != 0) { atomic_read 287 drivers/target/target_core_tpg.c while (atomic_read(&nacl->acl_pr_ref_count) != 0) atomic_read 526 drivers/target/target_core_tpg.c while (atomic_read(&se_tpg->tpg_pr_ref_count) != 0) atomic_read 2025 drivers/target/target_core_transport.c if (!atomic_read(&dev->simple_cmds)) atomic_read 2036 drivers/target/target_core_transport.c if (atomic_read(&dev->dev_ordered_sync) == 0) atomic_read 2280 drivers/target/target_core_transport.c if (atomic_read(&cmd->se_dev->dev_qf_count) != 0) atomic_read 224 drivers/target/target_core_user.c if (atomic_read(&global_db_count) > tcmu_global_max_blocks) atomic_read 1271 drivers/target/target_core_user.c if (atomic_read(&global_db_count) > atomic_read 2641 drivers/target/target_core_user.c if (atomic_read(&global_db_count) <= tcmu_global_max_blocks) atomic_read 2694 drivers/target/target_core_user.c if (atomic_read(&global_db_count) > tcmu_global_max_blocks) atomic_read 312 drivers/thermal/intel/intel_powerclamp.c atomic_read(&idle_wakeup_counter) > atomic_read 357 drivers/thermal/intel/intel_powerclamp.c reduce_irq = atomic_read(&idle_wakeup_counter) >= atomic_read 479 drivers/thermal/thermal_core.c if (atomic_read(&in_suspend)) atomic_read 493 drivers/tty/mips_ejtag_fdc.c atomic_read(&priv->xmit_total) || atomic_read 596 drivers/tty/rocket.c if (atomic_read(&rp_num_ports_open)) atomic_read 904 drivers/tty/rocket.c atomic_read(&rp_num_ports_open)); atomic_read 1035 drivers/tty/rocket.c atomic_read(&rp_num_ports_open)); atomic_read 293 drivers/tty/serial/atmel_serial.c if (!atomic_read(&atmel_port->tasklet_shutdown)) atomic_read 577 drivers/tty/serial/atmel_serial.c if (!atomic_read(&atmel_port->tasklet_shutdown)) atomic_read 1282 drivers/tty/serial/atmel_serial.c if (!atomic_read(&atmel_port->tasklet_shutdown)) { atomic_read 138 drivers/tty/serial/kgdb_nmi.c if (atomic_read(&kgdb_nmi_num_readers)) { atomic_read 199 drivers/tty/serial/kgdb_nmi.c if (likely(!atomic_read(&kgdb_nmi_num_readers) || atomic_read 2952 drivers/tty/serial/serial_core.c wait_event(state->remove_wait, !atomic_read(&state->refcount)); atomic_read 93 drivers/tty/tty_buffer.c int space = port->buf.mem_limit - atomic_read(&port->buf.mem_used); atomic_read 173 drivers/tty/tty_buffer.c if (atomic_read(&port->buf.mem_used) > port->buf.mem_limit) atomic_read 513 drivers/tty/tty_buffer.c if (atomic_read(&buf->priority)) atomic_read 259 drivers/uio/uio.c return sprintf(buf, "%u\n", (unsigned int)atomic_read(&idev->event)); atomic_read 487 drivers/uio/uio.c listener->event_count = atomic_read(&idev->event); atomic_read 558 drivers/uio/uio.c if (listener->event_count != atomic_read(&idev->event)) atomic_read 588 drivers/uio/uio.c event_count = atomic_read(&idev->event); atomic_read 745 drivers/usb/atm/usbatm.c atomic_read(&atm_dev->stats.aal5.tx), atomic_read 746 drivers/usb/atm/usbatm.c atomic_read(&atm_dev->stats.aal5.tx_err), atomic_read 747 drivers/usb/atm/usbatm.c atomic_read(&atm_dev->stats.aal5.rx), atomic_read 748 drivers/usb/atm/usbatm.c atomic_read(&atm_dev->stats.aal5.rx_err), atomic_read 749 drivers/usb/atm/usbatm.c atomic_read(&atm_dev->stats.aal5.rx_drop)); atomic_read 532 drivers/usb/class/usbtmc.c atomic_read(&data->iin_data_valid) != 0, atomic_read 593 drivers/usb/class/usbtmc.c atomic_read(&file_data->srq_asserted) != 0 || atomic_read 594 drivers/usb/class/usbtmc.c atomic_read(&file_data->closing), atomic_read 600 drivers/usb/class/usbtmc.c if (atomic_read(&file_data->closing) || data->zombie) atomic_read 2188 drivers/usb/class/usbtmc.c if (atomic_read(&file_data->srq_asserted)) atomic_read 632 drivers/usb/core/devices.c event_count = atomic_read(&device_event.count); atomic_read 1573 drivers/usb/core/driver.c __func__, atomic_read(&udev->dev.power.usage_count), atomic_read 1605 drivers/usb/core/driver.c __func__, atomic_read(&udev->dev.power.usage_count), atomic_read 1635 drivers/usb/core/driver.c __func__, atomic_read(&intf->dev.power.usage_count), atomic_read 1663 drivers/usb/core/driver.c __func__, atomic_read(&intf->dev.power.usage_count), atomic_read 1713 drivers/usb/core/driver.c __func__, atomic_read(&intf->dev.power.usage_count), atomic_read 1746 drivers/usb/core/driver.c __func__, atomic_read(&intf->dev.power.usage_count), atomic_read 1796 drivers/usb/core/driver.c if (atomic_read(&intf->dev.power.usage_count) > 0) atomic_read 1151 drivers/usb/core/hcd.c if (unlikely(atomic_read(&urb->reject))) { atomic_read 1564 drivers/usb/core/hcd.c if (atomic_read(&urb->reject)) atomic_read 1613 drivers/usb/core/hcd.c if (atomic_read(&urb->use_count) > 0) { atomic_read 1658 drivers/usb/core/hcd.c if (unlikely(atomic_read(&urb->reject))) atomic_read 297 drivers/usb/core/sysfs.c return sprintf(buf, "%d\n", atomic_read(&udev->urbnum)); atomic_read 144 drivers/usb/core/urb.c return atomic_read(&anchor->suspend_wakeups) == 0 && atomic_read 696 drivers/usb/core/urb.c wait_event(usb_kill_urb_queue, atomic_read(&urb->use_count) == 0); atomic_read 740 drivers/usb/core/urb.c wait_event(usb_kill_urb_queue, atomic_read(&urb->use_count) == 0); atomic_read 384 drivers/usb/gadget/function/f_ecm.c if (atomic_read(&ecm->notify_count)) atomic_read 913 drivers/usb/gadget/function/f_ecm.c if (atomic_read(&ecm->notify_count)) { atomic_read 1749 drivers/usb/gadget/function/f_fs.c if (atomic_read(&ffs->opened) < 0) { atomic_read 551 drivers/usb/gadget/function/f_ncm.c if (atomic_read(&ncm->notify_count)) atomic_read 1656 drivers/usb/gadget/function/f_ncm.c if (atomic_read(&ncm->notify_count)) { atomic_read 439 drivers/usb/gadget/function/f_tcm.c luns = atomic_read(&fu->tpg->tpg_port_count); atomic_read 1615 drivers/usb/gadget/function/f_tcm.c if (atomic_read(&tpg->tpg_port_count)) { atomic_read 1618 drivers/usb/gadget/function/f_tcm.c pr_err(MSG, atomic_read(&tpg->tpg_port_count)); atomic_read 2472 drivers/usb/gadget/udc/lpc32xx_udc.c if (atomic_read(&udc->enabled_ep_cnt)) atomic_read 2474 drivers/usb/gadget/udc/lpc32xx_udc.c (atomic_read(&udc->enabled_ep_cnt) == 0)); atomic_read 2945 drivers/usb/gadget/udc/lpc32xx_udc.c if (atomic_read(&udc->enabled_ep_cnt)) atomic_read 2947 drivers/usb/gadget/udc/lpc32xx_udc.c (atomic_read(&udc->enabled_ep_cnt) == 0)); atomic_read 487 drivers/usb/host/isp116x-hcd.c if (atomic_read(&isp116x->atl_finishing)) atomic_read 749 drivers/usb/host/isp1362-hcd.c if (atomic_read(&epq->finishing)) { atomic_read 801 drivers/usb/host/isp1362-hcd.c if (atomic_read(&epq->finishing)) { atomic_read 868 drivers/usb/host/isp1362-hcd.c if (atomic_read(&epq->finishing)) { atomic_read 178 drivers/usb/misc/iowarrior.c intr_idx = atomic_read(&dev->intr_idx); atomic_read 181 drivers/usb/misc/iowarrior.c read_idx = atomic_read(&dev->read_idx); atomic_read 268 drivers/usb/misc/iowarrior.c read_idx = atomic_read(&dev->read_idx); atomic_read 269 drivers/usb/misc/iowarrior.c intr_idx = atomic_read(&dev->intr_idx); atomic_read 332 drivers/usb/misc/iowarrior.c } while (atomic_read(&dev->overflow_flag)); atomic_read 393 drivers/usb/misc/iowarrior.c if (atomic_read(&dev->write_busy) == MAX_WRITES_IN_FLIGHT) { atomic_read 400 drivers/usb/misc/iowarrior.c (!dev->present || (atomic_read (&dev-> write_busy) < MAX_WRITES_IN_FLIGHT))); atomic_read 448 drivers/usb/misc/iowarrior.c retval, atomic_read(&dev->write_busy)); atomic_read 701 drivers/usb/misc/iowarrior.c if (atomic_read(&dev->write_busy) < MAX_WRITES_IN_FLIGHT) atomic_read 853 drivers/usb/serial/io_edgeport.c __func__, urb, atomic_read(&CmdUrbs)); atomic_read 2260 drivers/usb/serial/io_edgeport.c __func__, urb, atomic_read(&CmdUrbs)); atomic_read 566 drivers/usb/serial/mos7720.c status = atomic_read(&mos_parport->shadowDSR) & 0xf8; atomic_read 768 drivers/usb/storage/realtek_cr.c atomic_read(&us->pusb_intf->dev.power.usage_count)); atomic_read 770 drivers/usb/storage/realtek_cr.c if (atomic_read(&us->pusb_intf->dev.power.usage_count) > 0) { atomic_read 777 drivers/usb/storage/realtek_cr.c atomic_read(&us->pusb_intf->dev.power.usage_count)); atomic_read 810 drivers/usb/storage/realtek_cr.c atomic_read(&us->pusb_intf->dev.power.usage_count)); atomic_read 812 drivers/usb/storage/realtek_cr.c if (atomic_read(&us->pusb_intf->dev.power.usage_count) <= 0) { atomic_read 71 drivers/usb/usbip/vhci_rx.c atomic_read(&vhci_hcd->seqnum)); atomic_read 701 drivers/vfio/vfio.c if (!atomic_read(&group->container_users)) atomic_read 714 drivers/vfio/vfio.c if (!atomic_read(&group->container_users)) atomic_read 1355 drivers/vfio/vfio.c if (atomic_read(&group->container_users)) atomic_read 1436 drivers/vfio/vfio.c if (0 == atomic_read(&group->container_users) || atomic_read 260 drivers/vhost/net.c wait_event(ubufs->wait, !atomic_read(&ubufs->refcount)); atomic_read 382 drivers/vhost/vsock.c val = atomic_read(&vsock->queued_replies); atomic_read 193 drivers/video/fbdev/arcfb.c int count = atomic_read(&par->ref_count); atomic_read 227 drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c if (atomic_read(&rg->map_count)) { atomic_read 664 drivers/video/fbdev/omap2/omapfb/omapfb-main.c WARN_ON(!atomic_read(&ofbi->region->lock_count)); atomic_read 847 drivers/video/fbdev/omap2/omapfb/omapfb-main.c WARN_ON(!atomic_read(&ofbi->region->lock_count)); atomic_read 941 drivers/video/fbdev/omap2/omapfb/omapfb-main.c WARN_ON(!atomic_read(&ofbi->region->lock_count)); atomic_read 1312 drivers/video/fbdev/omap2/omapfb/omapfb-main.c WARN_ON(atomic_read(&rg->map_count)); atomic_read 449 drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c if (atomic_read(&rg->map_count)) { atomic_read 513 drivers/video/fbdev/ps3fb.c if (atomic_read(&ps3fb.ext_flip)) { atomic_read 926 drivers/video/fbdev/ps3fb.c !atomic_read(&ps3fb.ext_flip)) { atomic_read 860 drivers/video/fbdev/smscufx.c if (!atomic_read(&dev->usb_active)) atomic_read 965 drivers/video/fbdev/smscufx.c if (!atomic_read(&dev->usb_active)) atomic_read 991 drivers/video/fbdev/smscufx.c if (!atomic_read(&dev->usb_active)) atomic_read 288 drivers/video/fbdev/udlfb.c if (!atomic_read(&dlfb->usb_active)) atomic_read 619 drivers/video/fbdev/udlfb.c if (!atomic_read(&dlfb->usb_active)) { atomic_read 799 drivers/video/fbdev/udlfb.c if (!atomic_read(&dlfb->usb_active)) atomic_read 877 drivers/video/fbdev/udlfb.c if (!atomic_read(&dlfb->usb_active)) atomic_read 1430 drivers/video/fbdev/udlfb.c atomic_read(&dlfb->bytes_rendered)); atomic_read 1438 drivers/video/fbdev/udlfb.c atomic_read(&dlfb->bytes_identical)); atomic_read 1446 drivers/video/fbdev/udlfb.c atomic_read(&dlfb->bytes_sent)); atomic_read 1454 drivers/video/fbdev/udlfb.c atomic_read(&dlfb->cpu_kcycles_used)); atomic_read 1166 drivers/video/fbdev/uvesafb.c int cnt = atomic_read(&par->ref_count); atomic_read 1187 drivers/video/fbdev/uvesafb.c int cnt = atomic_read(&par->ref_count); atomic_read 75 drivers/w1/slaves/w1_therm.c refcnt = atomic_read(THERM_REFCNT(sl->family_data)); atomic_read 73 drivers/w1/w1_family.c while (atomic_read(&fent->refcnt)) { atomic_read 75 drivers/w1/w1_family.c fent->fid, atomic_read(&fent->refcnt)); atomic_read 198 drivers/w1/w1_int.c while (atomic_read(&dev->refcnt)) { atomic_read 200 drivers/w1/w1_int.c dev->name, atomic_read(&dev->refcnt)); atomic_read 249 drivers/watchdog/pcwd_usb.c if (atomic_read(&usb_pcwd->cmd_received)) atomic_read 231 drivers/xen/pvcalls-back.c while (atomic_read(&map->io) > 0) { atomic_read 232 drivers/xen/pvcalls-back.c if (atomic_read(&map->release) > 0) { atomic_read 237 drivers/xen/pvcalls-back.c if (atomic_read(&map->read) > 0) atomic_read 239 drivers/xen/pvcalls-back.c if (atomic_read(&map->write) > 0) atomic_read 1054 drivers/xen/pvcalls-front.c while (atomic_read(&map->refcount) > 1) atomic_read 1062 drivers/xen/pvcalls-front.c while (atomic_read(&map->refcount) > 1) atomic_read 1106 drivers/xen/pvcalls-front.c while (atomic_read(&pvcalls_refcount) > 0) atomic_read 586 drivers/xen/xen-scsiback.c atomic_read(&info->nr_unreplied_reqs) == 0); atomic_read 25 fs/affs/file.c inode->i_ino, atomic_read(&AFFS_I(inode)->i_opencnt)); atomic_read 34 fs/affs/file.c inode->i_ino, atomic_read(&AFFS_I(inode)->i_opencnt)); atomic_read 482 fs/afs/cell.c ASSERTCMP(atomic_read(&cell->usage), ==, 0); atomic_read 691 fs/afs/cell.c if (atomic_read(&cell->usage) > 1) { atomic_read 704 fs/afs/cell.c if (atomic_read(&cell->usage) > 1) atomic_read 738 fs/afs/cell.c _leave(" [destruct %d]", atomic_read(&net->cells_outstanding)); atomic_read 775 fs/afs/cell.c usage = atomic_read(&cell->usage); atomic_read 830 fs/afs/cell.c _leave(" [%d]", atomic_read(&net->cells_outstanding)); atomic_read 857 fs/afs/cell.c !atomic_read(&net->cells_outstanding)); atomic_read 271 fs/afs/cmservice.c trace_afs_server(call->server, atomic_read(&call->server->usage), atomic_read 194 fs/afs/dynroot.c if (!sb || atomic_read(&sb->s_active) == 0) atomic_read 223 fs/afs/dynroot.c if (!sb || atomic_read(&sb->s_active) == 0) atomic_read 50 fs/afs/proc.c atomic_read(&cell->usage), atomic_read 221 fs/afs/proc.c atomic_read(&vol->usage), vol->vid, atomic_read 388 fs/afs/proc.c atomic_read(&server->usage), atomic_read 578 fs/afs/proc.c atomic_read(&net->n_lookup), atomic_read 579 fs/afs/proc.c atomic_read(&net->n_reval), atomic_read 580 fs/afs/proc.c atomic_read(&net->n_inval), atomic_read 581 fs/afs/proc.c atomic_read(&net->n_relpg)); atomic_read 584 fs/afs/proc.c atomic_read(&net->n_read_dir)); atomic_read 587 fs/afs/proc.c atomic_read(&net->n_dir_cr), atomic_read 588 fs/afs/proc.c atomic_read(&net->n_dir_rm)); atomic_read 591 fs/afs/proc.c atomic_read(&net->n_fetches), atomic_read 594 fs/afs/proc.c atomic_read(&net->n_stores), atomic_read 120 fs/afs/rxrpc.c _debug("outstanding %u", atomic_read(&net->nr_outstanding_calls)); atomic_read 122 fs/afs/rxrpc.c !atomic_read(&net->nr_outstanding_calls)); atomic_read 169 fs/afs/rxrpc.c int o = atomic_read(&net->nr_outstanding_calls); atomic_read 207 fs/afs/rxrpc.c atomic_read(&call->net->nr_outstanding_calls), atomic_read 377 fs/afs/rxrpc.c atomic_read(&call->net->nr_outstanding_calls)); atomic_read 709 fs/afs/rxrpc.c atomic_read(&call->net->nr_outstanding_calls), atomic_read 295 fs/afs/server.c _leave(" = %p{%d}", server, atomic_read(&server->usage)); atomic_read 364 fs/afs/server.c trace_afs_server(server, atomic_read(&server->usage), atomic_read 382 fs/afs/server.c trace_afs_server(server, atomic_read(&server->usage), atomic_read 389 fs/afs/server.c atomic_read(&server->probe_outstanding) == 0); atomic_read 391 fs/afs/server.c trace_afs_server(server, atomic_read(&server->usage), atomic_read 456 fs/afs/server.c int usage = atomic_read(&server->usage); atomic_read 498 fs/afs/server.c _leave(" [%d]", atomic_read(&net->servers_outstanding)); atomic_read 522 fs/afs/server.c !atomic_read(&net->servers_outstanding)); atomic_read 535 fs/afs/server.c trace_afs_server(server, atomic_read(&server->usage), afs_server_trace_update); atomic_read 143 fs/afs/super.c if (atomic_read(&afs_count_active_inodes) != 0) { atomic_read 145 fs/afs/super.c atomic_read(&afs_count_active_inodes)); atomic_read 343 fs/aio.c if (!atomic_read(&ctx->dead)) { atomic_read 929 fs/aio.c int old, avail = atomic_read(&ctx->reqs_available); atomic_read 1262 fs/aio.c if (unlikely(atomic_read(&ctx->dead))) atomic_read 214 fs/autofs/expire.c unsigned int ino_count = atomic_read(&ino->count); atomic_read 382 fs/autofs/expire.c ino_count = atomic_read(&ino->count) + 1; atomic_read 399 fs/autofs/expire.c ino_count = atomic_read(&ino->count) + 1; atomic_read 79 fs/btrfs/async-thread.c return atomic_read(&wq->normal->pending) > wq->normal->thresh * 2; atomic_read 204 fs/btrfs/async-thread.c pending = atomic_read(&wq->pending); atomic_read 319 fs/btrfs/block-group.c wait_var_event(&bg->nocow_writers, !atomic_read(&bg->nocow_writers)); atomic_read 356 fs/btrfs/block-group.c wait_var_event(&bg->reservations, !atomic_read(&bg->reservations)); atomic_read 1090 fs/btrfs/block-group.c remove_em = (atomic_read(&block_group->trimming) == 0); atomic_read 2604 fs/btrfs/block-group.c atomic_read(&cur_trans->num_writers) == 1); atomic_read 3145 fs/btrfs/block-group.c ASSERT(atomic_read(&block_group->count) == 1); atomic_read 916 fs/btrfs/compression.c if (atomic_read(total_ws) > cpus) { atomic_read 921 fs/btrfs/compression.c if (atomic_read(total_ws) > cpus && !*free_ws) atomic_read 952 fs/btrfs/compression.c if (atomic_read(total_ws) == 0) { atomic_read 1331 fs/btrfs/delayed-inode.c if (atomic_read(&delayed_root->items) < atomic_read 1398 fs/btrfs/delayed-inode.c int val = atomic_read(&delayed_root->items_seq); atomic_read 1403 fs/btrfs/delayed-inode.c if (atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND) atomic_read 1413 fs/btrfs/delayed-inode.c if ((atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND) || atomic_read 1417 fs/btrfs/delayed-inode.c if (atomic_read(&delayed_root->items) >= BTRFS_DELAYED_WRITEBACK) { atomic_read 1421 fs/btrfs/delayed-inode.c seq = atomic_read(&delayed_root->items_seq); atomic_read 56 fs/btrfs/delayed-ref.c atomic_read(&trans->transaction->delayed_refs.num_entries); atomic_read 871 fs/btrfs/disk-io.c if (atomic_read(&bi->sync_writers)) atomic_read 1006 fs/btrfs/disk-io.c BUG_ON(!atomic_read(&eb->refs)); atomic_read 4006 fs/btrfs/disk-io.c (atomic_read(&fs_info->defrag_running) == 0)); atomic_read 4276 fs/btrfs/disk-io.c if (atomic_read(&delayed_refs->num_entries) == 0) { atomic_read 4582 fs/btrfs/disk-io.c atomic_read(&t->num_writers) == 0); atomic_read 2180 fs/btrfs/extent-tree.c count = atomic_read(&delayed_refs->num_entries) * 2; atomic_read 5751 fs/btrfs/extent-tree.c if (atomic_read(&root->will_be_snapshotted)) atomic_read 5759 fs/btrfs/extent-tree.c if (atomic_read(&root->will_be_snapshotted)) { atomic_read 5775 fs/btrfs/extent-tree.c !atomic_read(&root->will_be_snapshotted)); atomic_read 81 fs/btrfs/extent_io.c eb->start, eb->len, atomic_read(&eb->refs), eb->bflags); atomic_read 4856 fs/btrfs/extent_io.c return (atomic_read(&eb->io_pages) || atomic_read 5048 fs/btrfs/extent_io.c refs = atomic_read(&eb->refs); atomic_read 5287 fs/btrfs/extent_io.c WARN_ON(atomic_read(&eb->refs) == 0); atomic_read 5326 fs/btrfs/extent_io.c refs = atomic_read(&eb->refs); atomic_read 5337 fs/btrfs/extent_io.c if (atomic_read(&eb->refs) == 2 && atomic_read 5358 fs/btrfs/extent_io.c if (atomic_read(&eb->refs) == 2 && !extent_buffer_under_io(eb) && atomic_read 5389 fs/btrfs/extent_io.c WARN_ON(atomic_read(&eb->refs) == 0); atomic_read 5403 fs/btrfs/extent_io.c WARN_ON(atomic_read(&eb->refs) == 0); atomic_read 6111 fs/btrfs/extent_io.c if (atomic_read(&eb->refs) != 1 || extent_buffer_under_io(eb)) { atomic_read 1512 fs/btrfs/inode.c if (!freespace_inode && atomic_read(&root->snapshot_force_cow)) atomic_read 2090 fs/btrfs/inode.c int async = !atomic_read(&BTRFS_I(inode)->sync_writers); atomic_read 3507 fs/btrfs/inode.c atomic_read(&fs_info->nr_delayed_iputs) == 0); atomic_read 4485 fs/btrfs/inode.c if (atomic_read(&inode->i_count) > 1) atomic_read 8465 fs/btrfs/inode.c async_submit = !atomic_read(&BTRFS_I(inode)->sync_writers); atomic_read 770 fs/btrfs/ioctl.c if (atomic_read(&root->nr_swapfiles)) { atomic_read 4537 fs/btrfs/ioctl.c if (atomic_read(&fs_info->balance_pause_req)) atomic_read 4539 fs/btrfs/ioctl.c if (atomic_read(&fs_info->balance_cancel_req)) atomic_read 41 fs/btrfs/locking.c WARN_ON(atomic_read(&eb->spinning_readers) == 0); atomic_read 57 fs/btrfs/locking.c BUG_ON(!atomic_read(&eb->read_locks)); atomic_read 210 fs/btrfs/locking.c if (eb->blocking_writers || atomic_read(&eb->blocking_readers)) atomic_read 214 fs/btrfs/locking.c if (eb->blocking_writers || atomic_read(&eb->blocking_readers)) { atomic_read 264 fs/btrfs/locking.c WARN_ON(atomic_read(&eb->blocking_readers) == 0); atomic_read 284 fs/btrfs/locking.c wait_event(eb->read_lock_wq, atomic_read(&eb->blocking_readers) == 0); atomic_read 287 fs/btrfs/locking.c if (atomic_read(&eb->blocking_readers) || eb->blocking_writers) { atomic_read 156 fs/btrfs/print-tree.c atomic_read(&eb->refs), eb->write_locks, atomic_read 157 fs/btrfs/print-tree.c atomic_read(&eb->read_locks), atomic_read 159 fs/btrfs/print-tree.c atomic_read(&eb->blocking_readers), atomic_read 161 fs/btrfs/print-tree.c atomic_read(&eb->spinning_readers), atomic_read 922 fs/btrfs/raid56.c if (atomic_read(&rbio->error) > max_errors) atomic_read 1333 fs/btrfs/raid56.c BUG_ON(atomic_read(&rbio->stripes_pending) == 0); atomic_read 1491 fs/btrfs/raid56.c if (atomic_read(&rbio->error) > rbio->bbio->max_errors) atomic_read 2042 fs/btrfs/raid56.c if (atomic_read(&rbio->error) > rbio->bbio->max_errors) atomic_read 2110 fs/btrfs/raid56.c if (atomic_read(&rbio->error) <= rbio->bbio->max_errors) { atomic_read 2532 fs/btrfs/raid56.c if (atomic_read(&rbio->error) > rbio->bbio->max_errors) atomic_read 783 fs/btrfs/reada.c if (atomic_read(&device->reada_in_flight) < atomic_read 807 fs/btrfs/reada.c if (atomic_read(&fs_info->reada_works_cnt) > atomic_read 843 fs/btrfs/reada.c atomic_read(&device->reada_in_flight)); atomic_read 976 fs/btrfs/reada.c while (atomic_read(&rc->elems)) { atomic_read 977 fs/btrfs/reada.c if (!atomic_read(&fs_info->reada_works_cnt)) atomic_read 979 fs/btrfs/reada.c wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0, atomic_read 981 fs/btrfs/reada.c dump_devs(fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0); atomic_read 984 fs/btrfs/reada.c dump_devs(fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0); atomic_read 996 fs/btrfs/reada.c while (atomic_read(&rc->elems)) { atomic_read 997 fs/btrfs/reada.c if (!atomic_read(&fs_info->reada_works_cnt)) atomic_read 999 fs/btrfs/reada.c wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0, atomic_read 279 fs/btrfs/scrub.c while (atomic_read(&fs_info->scrub_pause_req)) { atomic_read 282 fs/btrfs/scrub.c atomic_read(&fs_info->scrub_pause_req) == 0); atomic_read 3142 fs/btrfs/scrub.c atomic_read(&sctx->bios_in_flight) == 0); atomic_read 3182 fs/btrfs/scrub.c if (atomic_read(&fs_info->scrub_cancel_req) || atomic_read 3183 fs/btrfs/scrub.c atomic_read(&sctx->cancel_req)) { atomic_read 3190 fs/btrfs/scrub.c if (atomic_read(&fs_info->scrub_pause_req)) { atomic_read 3198 fs/btrfs/scrub.c atomic_read(&sctx->bios_in_flight) == 0); atomic_read 3649 fs/btrfs/scrub.c atomic_read(&sctx->bios_in_flight) == 0); atomic_read 3659 fs/btrfs/scrub.c atomic_read(&sctx->workers_pending) == 0); atomic_read 3740 fs/btrfs/scrub.c wait_event(sctx->list_wait, atomic_read(&sctx->bios_in_flight) == 0); atomic_read 3931 fs/btrfs/scrub.c wait_event(sctx->list_wait, atomic_read(&sctx->bios_in_flight) == 0); atomic_read 3935 fs/btrfs/scrub.c wait_event(sctx->list_wait, atomic_read(&sctx->workers_pending) == 0); atomic_read 3974 fs/btrfs/scrub.c while (atomic_read(&fs_info->scrubs_paused) != atomic_read 3975 fs/btrfs/scrub.c atomic_read(&fs_info->scrubs_running)) { atomic_read 3978 fs/btrfs/scrub.c atomic_read(&fs_info->scrubs_paused) == atomic_read 3979 fs/btrfs/scrub.c atomic_read(&fs_info->scrubs_running)); atomic_read 3994 fs/btrfs/scrub.c if (!atomic_read(&fs_info->scrubs_running)) { atomic_read 4000 fs/btrfs/scrub.c while (atomic_read(&fs_info->scrubs_running)) { atomic_read 4003 fs/btrfs/scrub.c atomic_read(&fs_info->scrubs_running) == 0); atomic_read 409 fs/btrfs/space-info.c async_pages = atomic_read(&fs_info->async_delalloc_pages); atomic_read 424 fs/btrfs/space-info.c atomic_read(&fs_info->async_delalloc_pages) <= atomic_read 1699 fs/btrfs/super.c (atomic_read(&fs_info->defrag_running) == 0)); atomic_read 137 fs/btrfs/transaction.c return atomic_read(&trans->num_extwriters); atomic_read 874 fs/btrfs/transaction.c WARN_ON(atomic_read(&cur_trans->num_writers) < 1); atomic_read 1857 fs/btrfs/transaction.c atomic_read(&cur_trans->num_writers) == 1); atomic_read 2119 fs/btrfs/transaction.c atomic_read(&cur_trans->num_writers) == 1); atomic_read 2968 fs/btrfs/tree-log.c atomic_read(&root->log_commit[index]))) atomic_read 2985 fs/btrfs/tree-log.c if (!atomic_read(&root->log_writers)) atomic_read 3059 fs/btrfs/tree-log.c if (atomic_read(&root->log_commit[index1])) { atomic_read 3068 fs/btrfs/tree-log.c if (atomic_read(&root->log_commit[(index1 + 1) % 2])) atomic_read 3072 fs/btrfs/tree-log.c int batch = atomic_read(&root->log_batch); atomic_read 3081 fs/btrfs/tree-log.c if (batch == atomic_read(&root->log_batch)) atomic_read 3189 fs/btrfs/tree-log.c if (atomic_read(&log_root_tree->log_commit[index2])) { atomic_read 3202 fs/btrfs/tree-log.c if (atomic_read(&log_root_tree->log_commit[(index2 + 1) % 2])) { atomic_read 620 fs/btrfs/volumes.c BUG_ON(atomic_read(&cur->__bi_cnt) == 0); atomic_read 3732 fs/btrfs/volumes.c if ((!counting && atomic_read(&fs_info->balance_pause_req)) || atomic_read 3733 fs/btrfs/volumes.c atomic_read(&fs_info->balance_cancel_req)) { atomic_read 3903 fs/btrfs/volumes.c return atomic_read(&fs_info->balance_cancel_req) || atomic_read 3904 fs/btrfs/volumes.c (atomic_read(&fs_info->balance_pause_req) == 0 && atomic_read 3905 fs/btrfs/volumes.c atomic_read(&fs_info->balance_cancel_req) == 0); atomic_read 4088 fs/btrfs/volumes.c atomic_read(&fs_info->balance_pause_req) || atomic_read 4089 fs/btrfs/volumes.c atomic_read(&fs_info->balance_cancel_req)) { atomic_read 4235 fs/btrfs/volumes.c if (ret == -ECANCELED && atomic_read(&fs_info->balance_pause_req)) atomic_read 4237 fs/btrfs/volumes.c else if (ret == -ECANCELED && atomic_read(&fs_info->balance_cancel_req)) atomic_read 6452 fs/btrfs/volumes.c if (atomic_read(&bbio->error) > bbio->max_errors) { atomic_read 6546 fs/btrfs/volumes.c if (atomic_read(&bbio->error) > bbio->max_errors) atomic_read 7505 fs/btrfs/volumes.c stats_cnt = atomic_read(&device->dev_stats_ccnt); atomic_read 515 fs/btrfs/volumes.h return atomic_read(dev->dev_stat_values + index); atomic_read 1141 fs/buffer.c if (atomic_read(&buf->b_count)) { atomic_read 3170 fs/buffer.c WARN_ON(atomic_read(&bh->b_count) < 1); atomic_read 3214 fs/buffer.c return atomic_read(&bh->b_count) | atomic_read 189 fs/cachefiles/interface.c _enter("{OBJ%x,%d}", _object->debug_id, atomic_read(&object->usage)); atomic_read 192 fs/cachefiles/interface.c ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); atomic_read 274 fs/cachefiles/interface.c object->fscache.debug_id, atomic_read(&object->usage)); atomic_read 280 fs/cachefiles/interface.c ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); atomic_read 334 fs/cachefiles/interface.c object->fscache.debug_id, atomic_read(&object->usage)); atomic_read 337 fs/cachefiles/interface.c ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); atomic_read 34 fs/cachefiles/proc.c x = atomic_read(&cachefiles_lookup_histogram[index]); atomic_read 35 fs/cachefiles/proc.c y = atomic_read(&cachefiles_mkdir_histogram[index]); atomic_read 36 fs/cachefiles/proc.c z = atomic_read(&cachefiles_create_histogram[index]); atomic_read 702 fs/cachefiles/rdwr.c object->fscache.debug_id, atomic_read(&op->op.usage), atomic_read 2764 fs/ceph/caps.c flags = atomic_read(&fi->num_locks) ? CHECK_FILELOCK : 0; atomic_read 345 fs/ceph/dir.c int shared_gen = atomic_read(&ci->i_shared_gen); atomic_read 766 fs/ceph/dir.c di->lease_shared_gen = atomic_read(&ci->i_shared_gen); atomic_read 1507 fs/ceph/dir.c if (atomic_read(&ci->i_shared_gen) == di->lease_shared_gen && atomic_read 1531 fs/ceph/dir.c shared_gen = atomic_read(&ci->i_shared_gen); atomic_read 1545 fs/ceph/dir.c dir, (unsigned)atomic_read(&ci->i_shared_gen), dentry, valid); atomic_read 1725 fs/ceph/dir.c di->lease_shared_gen == atomic_read(&dir_ci->i_shared_gen)) atomic_read 1060 fs/ceph/inode.c di->lease_shared_gen = atomic_read(&ceph_inode(dir)->i_shared_gen); atomic_read 1671 fs/ceph/inode.c atomic_read(&ci->i_shared_gen)) { atomic_read 1450 fs/ceph/mds_client.c if (atomic_read(&ci->i_filelock_ref) > 0) { atomic_read 1752 fs/ceph/mds_client.c if (atomic_read(&ci->i_filelock_ref) > 0) atomic_read 1778 fs/ceph/mds_client.c count = atomic_read(&inode->i_count); atomic_read 4334 fs/ceph/mds_client.c return atomic_read(&mdsc->num_sessions) <= skipped; atomic_read 69 fs/ceph/snap.c atomic_read(&realm->nref), atomic_read(&realm->nref)+1); atomic_read 202 fs/ceph/snap.c atomic_read(&realm->nref), atomic_read(&realm->nref)-1); atomic_read 214 fs/ceph/snap.c atomic_read(&realm->nref), atomic_read(&realm->nref)-1); atomic_read 1154 fs/ceph/snap.c if (WARN_ON_ONCE(atomic_read(&sm->ref))) { atomic_read 302 fs/cifs/cifs_debug.c atomic_read(&server->smbd_conn->send_credits), atomic_read 303 fs/cifs/cifs_debug.c atomic_read(&server->smbd_conn->receive_credits), atomic_read 307 fs/cifs/cifs_debug.c atomic_read(&server->smbd_conn->send_pending), atomic_read 308 fs/cifs/cifs_debug.c atomic_read(&server->smbd_conn->send_payload_pending)); atomic_read 319 fs/cifs/cifs_debug.c atomic_read(&server->smbd_conn->mr_ready_count), atomic_read 320 fs/cifs/cifs_debug.c atomic_read(&server->smbd_conn->mr_used_count)); atomic_read 370 fs/cifs/cifs_debug.c atomic_read(&server->in_send), atomic_read 371 fs/cifs/cifs_debug.c atomic_read(&server->num_waiters)); atomic_read 517 fs/cifs/cifs_debug.c atomic_read(&totBufAllocCount), atomic_read 518 fs/cifs/cifs_debug.c atomic_read(&totSmBufAllocCount)); atomic_read 521 fs/cifs/cifs_debug.c seq_printf(m, "Operations (MIDs): %d\n", atomic_read(&midCount)); atomic_read 543 fs/cifs/cifs_debug.c atomic_read(&server->num_cmds[j]), atomic_read 548 fs/cifs/cifs_debug.c if (atomic_read(&server->smb2slowcmd[j])) atomic_read 550 fs/cifs/cifs_debug.c atomic_read(&server->smb2slowcmd[j]), atomic_read 565 fs/cifs/cifs_debug.c atomic_read(&tcon->num_smbs_sent)); atomic_read 1279 fs/cifs/connect.c atomic_read(&midCount)); atomic_read 5492 fs/cifs/connect.c atomic_read(&tlink->tl_count) != 0 || atomic_read 645 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_oplock_brks)); atomic_read 647 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_reads), atomic_read 650 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_writes), atomic_read 653 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_flushes)); atomic_read 655 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_locks), atomic_read 656 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_hardlinks), atomic_read 657 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_symlinks)); atomic_read 659 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_opens), atomic_read 660 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_closes), atomic_read 661 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_deletes)); atomic_read 663 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_posixopens), atomic_read 664 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_posixmkdirs)); atomic_read 666 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_mkdirs), atomic_read 667 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_rmdirs)); atomic_read 669 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_renames), atomic_read 670 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_t2renames)); atomic_read 672 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_ffirst), atomic_read 673 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_fnext), atomic_read 674 fs/cifs/smb1ops.c atomic_read(&tcon->stats.cifs_stats.num_fclose)); atomic_read 1292 fs/cifs/smb2ops.c atomic_read(&tcon->num_local_opens), atomic_read 1293 fs/cifs/smb2ops.c atomic_read(&tcon->num_remote_opens)); atomic_read 1295 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_TREE_CONNECT_HE]), atomic_read 1296 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_TREE_CONNECT_HE])); atomic_read 1298 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_TREE_DISCONNECT_HE]), atomic_read 1299 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_TREE_DISCONNECT_HE])); atomic_read 1301 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_CREATE_HE]), atomic_read 1302 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_CREATE_HE])); atomic_read 1304 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_CLOSE_HE]), atomic_read 1305 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_CLOSE_HE])); atomic_read 1307 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_FLUSH_HE]), atomic_read 1308 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_FLUSH_HE])); atomic_read 1310 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_READ_HE]), atomic_read 1311 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_READ_HE])); atomic_read 1313 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_WRITE_HE]), atomic_read 1314 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_WRITE_HE])); atomic_read 1316 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_LOCK_HE]), atomic_read 1317 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_LOCK_HE])); atomic_read 1319 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_IOCTL_HE]), atomic_read 1320 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_IOCTL_HE])); atomic_read 1322 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_QUERY_DIRECTORY_HE]), atomic_read 1323 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_QUERY_DIRECTORY_HE])); atomic_read 1325 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_CHANGE_NOTIFY_HE]), atomic_read 1326 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_CHANGE_NOTIFY_HE])); atomic_read 1328 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_QUERY_INFO_HE]), atomic_read 1329 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_QUERY_INFO_HE])); atomic_read 1331 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_SET_INFO_HE]), atomic_read 1332 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_SET_INFO_HE])); atomic_read 1334 fs/cifs/smb2ops.c atomic_read(&sent[SMB2_OPLOCK_BREAK_HE]), atomic_read 1335 fs/cifs/smb2ops.c atomic_read(&failed[SMB2_OPLOCK_BREAK_HE])); atomic_read 401 fs/cifs/smbdirect.c if (atomic_read(&info->receive_credits) < atomic_read 423 fs/cifs/smbdirect.c atomic_read(&info->receive_credits)) { atomic_read 471 fs/cifs/smbdirect.c if (atomic_read(&info->send_credits)) atomic_read 840 fs/cifs/smbdirect.c atomic_read(&info->send_credits) > 0 || atomic_read 1422 fs/cifs/smbdirect.c atomic_read(&info->send_pending) == 0); atomic_read 1424 fs/cifs/smbdirect.c atomic_read(&info->send_payload_pending) == 0); atomic_read 1457 fs/cifs/smbdirect.c while (atomic_read(&info->mr_used_count)) { atomic_read 2239 fs/cifs/smbdirect.c atomic_read(&info->send_payload_pending) == 0); atomic_read 2405 fs/cifs/smbdirect.c atomic_read(&info->mr_ready_count) || atomic_read 119 fs/cifs/transport.c if (atomic_read(&server->num_cmds[smb_cmd]) == 0) { atomic_read 36 fs/coda/cache.c cii->c_cached_epoch = atomic_read(&permission_epoch); atomic_read 50 fs/coda/cache.c cii->c_cached_epoch = atomic_read(&permission_epoch) - 1; atomic_read 70 fs/coda/cache.c cii->c_cached_epoch == atomic_read(&permission_epoch); atomic_read 147 fs/configfs/configfs_internal.h WARN_ON(!atomic_read(&sd->s_count)); atomic_read 155 fs/configfs/configfs_internal.h WARN_ON(!atomic_read(&sd->s_count)); atomic_read 381 fs/coredump.c if (atomic_read(&mm->mm_users) == nr + 1) atomic_read 780 fs/debugfs/file.c *val = atomic_read((atomic_t *)data); atomic_read 920 fs/dlm/user.c return atomic_read(&dlm_monitor_opened) ? 1 : 0; atomic_read 492 fs/ecryptfs/miscdev.c BUG_ON(atomic_read(&ecryptfs_num_miscdev_opens) != 0); atomic_read 157 fs/erofs/internal.h if (val != atomic_read(&grp->refcount)) { atomic_read 172 fs/erofs/internal.h int v = atomic_read(&grp->refcount); atomic_read 94 fs/erofs/utils.c if (atomic_read(&grp->refcount) != 1) { atomic_read 1326 fs/erofs/zdata.c !atomic_read(&io[JQ_SUBMIT].pending_bios)); atomic_read 132 fs/erofs/zdata.h return atomic_read(u.o) >> Z_EROFS_ONLINEPAGE_INDEX_SHIFT; atomic_read 185 fs/erofs/zdata.h erofs_dbg("%s, page %p value %x", __func__, page, atomic_read(u.o)); atomic_read 1738 fs/exec.c atomic_read(¤t_user()->processes) > rlimit(RLIMIT_NPROC)) { atomic_read 225 fs/ext2/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(HDR(bh)->h_refcount)); atomic_read 311 fs/ext2/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(HDR(bh)->h_refcount)); atomic_read 448 fs/ext2/xattr.c atomic_read(&(bh->b_count)), atomic_read 812 fs/ext2/xattr.c ea_bdebug(bh, "b_count=%d", atomic_read(&(bh->b_count))); atomic_read 968 fs/ext2/xattr.c atomic_read(&(bh->b_count))); atomic_read 3072 fs/ext4/ext4.h return (atomic_read(&sbi->s_lock_busy) > EXT4_CONTENTION_THRESHOLD); atomic_read 1771 fs/ext4/extents.c atomic_read(&EXT4_I(inode)->i_unwritten) || atomic_read 94 fs/ext4/file.c (atomic_read(&inode->i_writecount) == 1) && atomic_read 111 fs/ext4/file.c wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_unwritten) == 0)); atomic_read 251 fs/ext4/ialloc.c if (atomic_read(&inode->i_count) > 1) { atomic_read 254 fs/ext4/ialloc.c atomic_read(&inode->i_count)); atomic_read 374 fs/ext4/ialloc.c stats->free_inodes = atomic_read(&fg->free_inodes); atomic_read 376 fs/ext4/ialloc.c stats->used_dirs = atomic_read(&fg->used_dirs); atomic_read 4243 fs/ext4/inode.c atomic_read(&page->_refcount) == 1, atomic_read 2752 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_allocated), atomic_read 2753 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_reqs), atomic_read 2754 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_success)); atomic_read 2758 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_ex_scanned), atomic_read 2759 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_goals), atomic_read 2760 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_2orders), atomic_read 2761 fs/ext4/mballoc.c atomic_read(&sbi->s_bal_breaks), atomic_read 2762 fs/ext4/mballoc.c atomic_read(&sbi->s_mb_lost_chunks)); atomic_read 2769 fs/ext4/mballoc.c atomic_read(&sbi->s_mb_preallocated), atomic_read 2770 fs/ext4/mballoc.c atomic_read(&sbi->s_mb_discarded)); atomic_read 3577 fs/ext4/mballoc.c BUG_ON(atomic_read(&pa->pa_count)); atomic_read 3941 fs/ext4/mballoc.c if (atomic_read(&pa->pa_count)) { atomic_read 4039 fs/ext4/mballoc.c if (atomic_read(&pa->pa_count)) { atomic_read 4306 fs/ext4/mballoc.c if (atomic_read(&pa->pa_count)) { atomic_read 5198 fs/ext4/mballoc.c minblocks >= atomic_read(&EXT4_SB(sb)->s_last_trim_minblks)) atomic_read 325 fs/ext4/sysfs.c atomic_read((atomic_t *) ptr)); atomic_read 533 fs/ext4/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(BHDR(bh)->h_refcount)); atomic_read 706 fs/ext4/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(BHDR(bh)->h_refcount)); atomic_read 1826 fs/ext4/xattr.c atomic_read(&(bs->bh->b_count)), atomic_read 2310 fs/f2fs/data.c if (atomic_read(&sbi->wb_sync_req[DATA]) && atomic_read 2455 fs/f2fs/data.c else if (atomic_read(&sbi->wb_sync_req[DATA])) atomic_read 45 fs/f2fs/debug.c si->ext_tree = atomic_read(&sbi->total_ext_tree); atomic_read 46 fs/f2fs/debug.c si->zombie_tree = atomic_read(&sbi->total_zombie_tree); atomic_read 47 fs/f2fs/debug.c si->ext_node = atomic_read(&sbi->total_ext_node); atomic_read 59 fs/f2fs/debug.c si->aw_cnt = atomic_read(&sbi->aw_cnt); atomic_read 60 fs/f2fs/debug.c si->vw_cnt = atomic_read(&sbi->vw_cnt); atomic_read 61 fs/f2fs/debug.c si->max_aw_cnt = atomic_read(&sbi->max_aw_cnt); atomic_read 62 fs/f2fs/debug.c si->max_vw_cnt = atomic_read(&sbi->max_vw_cnt); atomic_read 72 fs/f2fs/debug.c atomic_read(&SM_I(sbi)->fcc_info->issued_flush); atomic_read 74 fs/f2fs/debug.c atomic_read(&SM_I(sbi)->fcc_info->queued_flush); atomic_read 80 fs/f2fs/debug.c atomic_read(&SM_I(sbi)->dcc_info->issued_discard); atomic_read 82 fs/f2fs/debug.c atomic_read(&SM_I(sbi)->dcc_info->queued_discard); atomic_read 84 fs/f2fs/debug.c atomic_read(&SM_I(sbi)->dcc_info->discard_cmd_cnt); atomic_read 94 fs/f2fs/debug.c si->inline_xattr = atomic_read(&sbi->inline_xattr); atomic_read 95 fs/f2fs/debug.c si->inline_inode = atomic_read(&sbi->inline_inode); atomic_read 96 fs/f2fs/debug.c si->inline_dir = atomic_read(&sbi->inline_dir); atomic_read 138 fs/f2fs/debug.c si->meta_count[i] = atomic_read(&sbi->meta_count[i]); atomic_read 145 fs/f2fs/debug.c si->inplace_count = atomic_read(&sbi->inplace_count); atomic_read 250 fs/f2fs/debug.c atomic_read(&SM_I(sbi)->dcc_info->discard_cmd_cnt); atomic_read 263 fs/f2fs/debug.c si->cache_mem += atomic_read(&sbi->total_ext_tree) * atomic_read 265 fs/f2fs/debug.c si->cache_mem += atomic_read(&sbi->total_ext_node) * atomic_read 304 fs/f2fs/extent_cache.c unsigned int count = atomic_read(&et->node_cnt); atomic_read 314 fs/f2fs/extent_cache.c return count - atomic_read(&et->node_cnt); atomic_read 352 fs/f2fs/extent_cache.c if (atomic_read(&et->node_cnt)) atomic_read 641 fs/f2fs/extent_cache.c if (!atomic_read(&sbi->total_zombie_tree)) atomic_read 649 fs/f2fs/extent_cache.c if (atomic_read(&et->node_cnt)) { atomic_read 654 fs/f2fs/extent_cache.c f2fs_bug_on(sbi, atomic_read(&et->node_cnt)); atomic_read 713 fs/f2fs/extent_cache.c if (!et || !atomic_read(&et->node_cnt)) atomic_read 755 fs/f2fs/extent_cache.c atomic_read(&et->node_cnt)) { atomic_read 768 fs/f2fs/extent_cache.c f2fs_bug_on(sbi, atomic_read(&et->node_cnt)); atomic_read 1385 fs/f2fs/f2fs.h if (atomic_read(&ffi->inject_ops) >= ffi->inject_rate) { atomic_read 1909 fs/f2fs/f2fs.h return atomic_read(&sbi->nr_pages[count_type]); atomic_read 1914 fs/f2fs/f2fs.h return atomic_read(&F2FS_I(inode)->dirty_pages); atomic_read 2248 fs/f2fs/f2fs.h atomic_read(&SM_I(sbi)->dcc_info->queued_discard)) atomic_read 2252 fs/f2fs/f2fs.h atomic_read(&SM_I(sbi)->fcc_info->queued_flush)) atomic_read 3381 fs/f2fs/f2fs.h int cur = atomic_read(&F2FS_I_SB(inode)->aw_cnt); \ atomic_read 3382 fs/f2fs/f2fs.h int max = atomic_read(&F2FS_I_SB(inode)->max_aw_cnt); \ atomic_read 3392 fs/f2fs/f2fs.h int cur = atomic_read(&F2FS_I_SB(inode)->vw_cnt); \ atomic_read 3393 fs/f2fs/f2fs.h int max = atomic_read(&F2FS_I_SB(inode)->max_vw_cnt); \ atomic_read 1635 fs/f2fs/file.c atomic_read(&inode->i_writecount) != 1) atomic_read 84 fs/f2fs/node.c mem_size = (atomic_read(&sbi->total_ext_tree) * atomic_read 86 fs/f2fs/node.c atomic_read(&sbi->total_ext_node) * atomic_read 1834 fs/f2fs/node.c if (atomic_read(&sbi->wb_sync_req[NODE]) && atomic_read 1993 fs/f2fs/node.c else if (atomic_read(&sbi->wb_sync_req[NODE])) atomic_read 1687 fs/f2fs/segment.c f2fs_bug_on(sbi, atomic_read(&dcc->discard_cmd_cnt)); atomic_read 1715 fs/f2fs/segment.c if (atomic_read(&dcc->queued_discard)) atomic_read 2091 fs/f2fs/segment.c if (unlikely(atomic_read(&dcc->discard_cmd_cnt))) atomic_read 35 fs/f2fs/shrinker.c return atomic_read(&sbi->total_zombie_tree) + atomic_read 36 fs/f2fs/shrinker.c atomic_read(&sbi->total_ext_node); atomic_read 160 fs/file.c if (atomic_read(&files->count) > 1) atomic_read 770 fs/file.c if (atomic_read(&files->count) == 1) { atomic_read 208 fs/fs-writeback.c wait_event(*done->waitq, !atomic_read(&done->cnt)); atomic_read 500 fs/fs-writeback.c if (atomic_read(&isw_nr_in_flight) > WB_FRN_MAX_IN_FLIGHT) atomic_read 1002 fs/fs-writeback.c if (atomic_read(&isw_nr_in_flight)) { atomic_read 1534 fs/fs-writeback.c if (!atomic_read(&inode->i_count)) atomic_read 402 fs/fscache/cache.c atomic_read(&cache->object_count) == 0); atomic_read 38 fs/fscache/cookie.c atomic_read(&cookie->n_children), atomic_read 39 fs/fscache/cookie.c atomic_read(&cookie->n_active)); atomic_read 221 fs/fscache/cookie.c atomic_read(&cursor->usage)); atomic_read 720 fs/fscache/cookie.c ASSERTCMP(atomic_read(&cookie->n_active), >, 0); atomic_read 722 fs/fscache/cookie.c if (atomic_read(&cookie->n_children) != 0) { atomic_read 766 fs/fscache/cookie.c !atomic_read(&cookie->n_active)); atomic_read 808 fs/fscache/cookie.c atomic_read(&cookie->n_active), retire); atomic_read 824 fs/fscache/cookie.c ASSERTCMP(atomic_read(&cookie->parent->usage), >, 0); atomic_read 825 fs/fscache/cookie.c ASSERTCMP(atomic_read(&cookie->parent->n_children), >, 0); atomic_read 830 fs/fscache/cookie.c ASSERTCMP(atomic_read(&cookie->usage), >, 0); atomic_read 37 fs/fscache/histogram.c n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); atomic_read 38 fs/fscache/histogram.c n[1] = atomic_read(&fscache_ops_histogram[index]); atomic_read 39 fs/fscache/histogram.c n[2] = atomic_read(&fscache_objs_histogram[index]); atomic_read 40 fs/fscache/histogram.c n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); atomic_read 41 fs/fscache/histogram.c n[4] = atomic_read(&fscache_retrieval_histogram[index]); atomic_read 225 fs/fscache/object-list.c FILTER(atomic_read(&obj->n_reads), atomic_read 242 fs/fscache/object-list.c atomic_read(&obj->n_reads), atomic_read 63 fs/fscache/operation.c op->object->debug_id, op->debug_id, atomic_read(&op->usage)); atomic_read 68 fs/fscache/operation.c ASSERTCMP(atomic_read(&op->usage), >, 0); atomic_read 166 fs/fscache/operation.c ASSERTCMP(atomic_read(&op->usage), >, 0); atomic_read 248 fs/fscache/operation.c object->debug_id, op->debug_id, atomic_read(&op->usage)); atomic_read 253 fs/fscache/operation.c ASSERTCMP(atomic_read(&op->usage), >, 0); atomic_read 373 fs/fscache/operation.c ASSERTCMP(atomic_read(&op->usage), >, 0); atomic_read 500 fs/fscache/operation.c op->debug_id, atomic_read(&op->usage)); atomic_read 502 fs/fscache/operation.c ASSERTCMP(atomic_read(&op->usage), >, 0); atomic_read 592 fs/fscache/operation.c ASSERTCMP(atomic_read(&op->usage), ==, 0); atomic_read 622 fs/fscache/operation.c op->object->debug_id, op->debug_id, atomic_read(&op->usage)); atomic_read 290 fs/fscache/page.c atomic_read(&op->n_pages), ==, 0); atomic_read 799 fs/fscache/page.c _enter("{OP%x,%d}", op->op.debug_id, atomic_read(&op->op.usage)); atomic_read 142 fs/fscache/stats.c atomic_read(&fscache_n_cookie_index), atomic_read 143 fs/fscache/stats.c atomic_read(&fscache_n_cookie_data), atomic_read 144 fs/fscache/stats.c atomic_read(&fscache_n_cookie_special)); atomic_read 147 fs/fscache/stats.c atomic_read(&fscache_n_object_alloc), atomic_read 148 fs/fscache/stats.c atomic_read(&fscache_n_object_no_alloc), atomic_read 149 fs/fscache/stats.c atomic_read(&fscache_n_object_avail), atomic_read 150 fs/fscache/stats.c atomic_read(&fscache_n_object_dead)); atomic_read 152 fs/fscache/stats.c atomic_read(&fscache_n_checkaux_none), atomic_read 153 fs/fscache/stats.c atomic_read(&fscache_n_checkaux_okay), atomic_read 154 fs/fscache/stats.c atomic_read(&fscache_n_checkaux_update), atomic_read 155 fs/fscache/stats.c atomic_read(&fscache_n_checkaux_obsolete)); atomic_read 158 fs/fscache/stats.c atomic_read(&fscache_n_marks), atomic_read 159 fs/fscache/stats.c atomic_read(&fscache_n_uncaches)); atomic_read 163 fs/fscache/stats.c atomic_read(&fscache_n_acquires), atomic_read 164 fs/fscache/stats.c atomic_read(&fscache_n_acquires_null), atomic_read 165 fs/fscache/stats.c atomic_read(&fscache_n_acquires_no_cache), atomic_read 166 fs/fscache/stats.c atomic_read(&fscache_n_acquires_ok), atomic_read 167 fs/fscache/stats.c atomic_read(&fscache_n_acquires_nobufs), atomic_read 168 fs/fscache/stats.c atomic_read(&fscache_n_acquires_oom)); atomic_read 171 fs/fscache/stats.c atomic_read(&fscache_n_object_lookups), atomic_read 172 fs/fscache/stats.c atomic_read(&fscache_n_object_lookups_negative), atomic_read 173 fs/fscache/stats.c atomic_read(&fscache_n_object_lookups_positive), atomic_read 174 fs/fscache/stats.c atomic_read(&fscache_n_object_created), atomic_read 175 fs/fscache/stats.c atomic_read(&fscache_n_object_lookups_timed_out)); atomic_read 178 fs/fscache/stats.c atomic_read(&fscache_n_invalidates), atomic_read 179 fs/fscache/stats.c atomic_read(&fscache_n_invalidates_run)); atomic_read 182 fs/fscache/stats.c atomic_read(&fscache_n_updates), atomic_read 183 fs/fscache/stats.c atomic_read(&fscache_n_updates_null), atomic_read 184 fs/fscache/stats.c atomic_read(&fscache_n_updates_run)); atomic_read 187 fs/fscache/stats.c atomic_read(&fscache_n_relinquishes), atomic_read 188 fs/fscache/stats.c atomic_read(&fscache_n_relinquishes_null), atomic_read 189 fs/fscache/stats.c atomic_read(&fscache_n_relinquishes_waitcrt), atomic_read 190 fs/fscache/stats.c atomic_read(&fscache_n_relinquishes_retire)); atomic_read 193 fs/fscache/stats.c atomic_read(&fscache_n_attr_changed), atomic_read 194 fs/fscache/stats.c atomic_read(&fscache_n_attr_changed_ok), atomic_read 195 fs/fscache/stats.c atomic_read(&fscache_n_attr_changed_nobufs), atomic_read 196 fs/fscache/stats.c atomic_read(&fscache_n_attr_changed_nomem), atomic_read 197 fs/fscache/stats.c atomic_read(&fscache_n_attr_changed_calls)); atomic_read 200 fs/fscache/stats.c atomic_read(&fscache_n_allocs), atomic_read 201 fs/fscache/stats.c atomic_read(&fscache_n_allocs_ok), atomic_read 202 fs/fscache/stats.c atomic_read(&fscache_n_allocs_wait), atomic_read 203 fs/fscache/stats.c atomic_read(&fscache_n_allocs_nobufs), atomic_read 204 fs/fscache/stats.c atomic_read(&fscache_n_allocs_intr)); atomic_read 206 fs/fscache/stats.c atomic_read(&fscache_n_alloc_ops), atomic_read 207 fs/fscache/stats.c atomic_read(&fscache_n_alloc_op_waits), atomic_read 208 fs/fscache/stats.c atomic_read(&fscache_n_allocs_object_dead)); atomic_read 212 fs/fscache/stats.c atomic_read(&fscache_n_retrievals), atomic_read 213 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_ok), atomic_read 214 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_wait), atomic_read 215 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_nodata), atomic_read 216 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_nobufs), atomic_read 217 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_intr), atomic_read 218 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_nomem)); atomic_read 220 fs/fscache/stats.c atomic_read(&fscache_n_retrieval_ops), atomic_read 221 fs/fscache/stats.c atomic_read(&fscache_n_retrieval_op_waits), atomic_read 222 fs/fscache/stats.c atomic_read(&fscache_n_retrievals_object_dead)); atomic_read 225 fs/fscache/stats.c atomic_read(&fscache_n_stores), atomic_read 226 fs/fscache/stats.c atomic_read(&fscache_n_stores_ok), atomic_read 227 fs/fscache/stats.c atomic_read(&fscache_n_stores_again), atomic_read 228 fs/fscache/stats.c atomic_read(&fscache_n_stores_nobufs), atomic_read 229 fs/fscache/stats.c atomic_read(&fscache_n_stores_oom)); atomic_read 231 fs/fscache/stats.c atomic_read(&fscache_n_store_ops), atomic_read 232 fs/fscache/stats.c atomic_read(&fscache_n_store_calls), atomic_read 233 fs/fscache/stats.c atomic_read(&fscache_n_store_pages), atomic_read 234 fs/fscache/stats.c atomic_read(&fscache_n_store_radix_deletes), atomic_read 235 fs/fscache/stats.c atomic_read(&fscache_n_store_pages_over_limit)); atomic_read 238 fs/fscache/stats.c atomic_read(&fscache_n_store_vmscan_not_storing), atomic_read 239 fs/fscache/stats.c atomic_read(&fscache_n_store_vmscan_gone), atomic_read 240 fs/fscache/stats.c atomic_read(&fscache_n_store_vmscan_busy), atomic_read 241 fs/fscache/stats.c atomic_read(&fscache_n_store_vmscan_cancelled), atomic_read 242 fs/fscache/stats.c atomic_read(&fscache_n_store_vmscan_wait)); atomic_read 245 fs/fscache/stats.c atomic_read(&fscache_n_op_pend), atomic_read 246 fs/fscache/stats.c atomic_read(&fscache_n_op_run), atomic_read 247 fs/fscache/stats.c atomic_read(&fscache_n_op_enqueue), atomic_read 248 fs/fscache/stats.c atomic_read(&fscache_n_op_cancelled), atomic_read 249 fs/fscache/stats.c atomic_read(&fscache_n_op_rejected)); atomic_read 251 fs/fscache/stats.c atomic_read(&fscache_n_op_initialised), atomic_read 252 fs/fscache/stats.c atomic_read(&fscache_n_op_deferred_release), atomic_read 253 fs/fscache/stats.c atomic_read(&fscache_n_op_release), atomic_read 254 fs/fscache/stats.c atomic_read(&fscache_n_op_gc)); atomic_read 257 fs/fscache/stats.c atomic_read(&fscache_n_cop_alloc_object), atomic_read 258 fs/fscache/stats.c atomic_read(&fscache_n_cop_lookup_object), atomic_read 259 fs/fscache/stats.c atomic_read(&fscache_n_cop_lookup_complete), atomic_read 260 fs/fscache/stats.c atomic_read(&fscache_n_cop_grab_object)); atomic_read 262 fs/fscache/stats.c atomic_read(&fscache_n_cop_invalidate_object), atomic_read 263 fs/fscache/stats.c atomic_read(&fscache_n_cop_update_object), atomic_read 264 fs/fscache/stats.c atomic_read(&fscache_n_cop_drop_object), atomic_read 265 fs/fscache/stats.c atomic_read(&fscache_n_cop_put_object), atomic_read 266 fs/fscache/stats.c atomic_read(&fscache_n_cop_attr_changed), atomic_read 267 fs/fscache/stats.c atomic_read(&fscache_n_cop_sync_cache)); atomic_read 269 fs/fscache/stats.c atomic_read(&fscache_n_cop_read_or_alloc_page), atomic_read 270 fs/fscache/stats.c atomic_read(&fscache_n_cop_read_or_alloc_pages), atomic_read 271 fs/fscache/stats.c atomic_read(&fscache_n_cop_allocate_page), atomic_read 272 fs/fscache/stats.c atomic_read(&fscache_n_cop_allocate_pages), atomic_read 273 fs/fscache/stats.c atomic_read(&fscache_n_cop_write_page), atomic_read 274 fs/fscache/stats.c atomic_read(&fscache_n_cop_uncache_page), atomic_read 275 fs/fscache/stats.c atomic_read(&fscache_n_cop_dissociate_pages)); atomic_read 277 fs/fscache/stats.c atomic_read(&fscache_n_cache_no_space_reject), atomic_read 278 fs/fscache/stats.c atomic_read(&fscache_n_cache_stale_objects), atomic_read 279 fs/fscache/stats.c atomic_read(&fscache_n_cache_retired_objects), atomic_read 280 fs/fscache/stats.c atomic_read(&fscache_n_cache_culled_objects)); atomic_read 59 fs/fuse/control.c value = atomic_read(&fc->num_waiting); atomic_read 583 fs/fuse/cuse.c return sprintf(buf, "%d\n", atomic_read(&cc->fc.num_waiting)); atomic_read 2157 fs/fuse/dev.c wait_event(fc->blocked_waitq, atomic_read(&fc->num_waiting) == 0); atomic_read 794 fs/gfs2/aops.c if (atomic_read(&bh->b_count)) atomic_read 1557 fs/gfs2/bmap.c if (isize_blks > atomic_read(&sdp->sd_log_thresh2)) atomic_read 1559 fs/gfs2/bmap.c atomic_read(&sdp->sd_log_thresh2); atomic_read 1575 fs/gfs2/bmap.c RES_QUOTA >= atomic_read(&sdp->sd_log_thresh2)) { atomic_read 403 fs/gfs2/file.c if (hint > atomic_read(&ip->i_sizehint)) atomic_read 140 fs/gfs2/glock.c BUG_ON(atomic_read(&gl->gl_revokes)); atomic_read 1625 fs/gfs2/glock.c return vfs_pressure_ratio(atomic_read(&lru_count)); atomic_read 1737 fs/gfs2/glock.c atomic_read(&sdp->sd_glock_disposal) == 0, atomic_read 1905 fs/gfs2/glock.c atomic_read(&gl->gl_ail_count), atomic_read 1906 fs/gfs2/glock.c atomic_read(&gl->gl_revokes), atomic_read 76 fs/gfs2/glops.c GLOCK_BUG_ON(gl, !fsync && atomic_read(&gl->gl_ail_count)); atomic_read 90 fs/gfs2/glops.c tr.tr_revokes = atomic_read(&gl->gl_ail_count); atomic_read 110 fs/gfs2/glops.c log_in_flight = atomic_read(&sdp->sd_log_in_flight); atomic_read 140 fs/gfs2/glops.c unsigned int revokes = atomic_read(&gl->gl_ail_count); atomic_read 219 fs/gfs2/glops.c gfs2_assert_withdraw(sdp, !atomic_read(&gl->gl_ail_count)); atomic_read 319 fs/gfs2/glops.c gfs2_assert_withdraw(gl->gl_name.ln_sbd, !atomic_read(&gl->gl_ail_count)); atomic_read 328 fs/gfs2/log.c gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= atomic_read 367 fs/gfs2/log.c free_blocks = atomic_read(&sdp->sd_log_blks_free); atomic_read 374 fs/gfs2/log.c if (atomic_read(&sdp->sd_log_blks_free) <= wanted) atomic_read 376 fs/gfs2/log.c free_blocks = atomic_read(&sdp->sd_log_blks_free); atomic_read 509 fs/gfs2/log.c gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= atomic_read 520 fs/gfs2/log.c if (atomic_read(&sdp->sd_log_in_flight)) { atomic_read 524 fs/gfs2/log.c if (atomic_read(&sdp->sd_log_in_flight)) atomic_read 526 fs/gfs2/log.c } while(atomic_read(&sdp->sd_log_in_flight)); atomic_read 766 fs/gfs2/log.c enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); atomic_read 795 fs/gfs2/log.c enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); atomic_read 921 fs/gfs2/log.c gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= atomic_read 947 fs/gfs2/log.c if (atomic_read(&sdp->sd_log_pinned) > atomic_read(&sdp->sd_log_thresh1) || atomic_read 948 fs/gfs2/log.c ((sdp->sd_jdesc->jd_blocks - atomic_read(&sdp->sd_log_blks_free)) > atomic_read 949 fs/gfs2/log.c atomic_read(&sdp->sd_log_thresh2))) atomic_read 978 fs/gfs2/log.c return (atomic_read(&sdp->sd_log_pinned) + atomic_read 979 fs/gfs2/log.c atomic_read(&sdp->sd_log_blks_needed) >= atomic_read 980 fs/gfs2/log.c atomic_read(&sdp->sd_log_thresh1)); atomic_read 985 fs/gfs2/log.c unsigned int used_blocks = sdp->sd_jdesc->jd_blocks - atomic_read(&sdp->sd_log_blks_free); atomic_read 990 fs/gfs2/log.c return used_blocks + atomic_read(&sdp->sd_log_blks_needed) >= atomic_read 991 fs/gfs2/log.c atomic_read(&sdp->sd_log_thresh2); atomic_read 725 fs/gfs2/ops_fstype.c trace_gfs2_log_blocks(sdp, atomic_read(&sdp->sd_log_blks_free)); atomic_read 553 fs/gfs2/quota.c if (ip->i_qadata && ((wcount == NULL) || (atomic_read(wcount) <= 1))) { atomic_read 1467 fs/gfs2/quota.c gfs2_assert_warn(sdp, !atomic_read(&sdp->sd_quota_count)); atomic_read 693 fs/gfs2/rgrp.c if ((wcount == NULL) || (atomic_read(wcount) <= 1)) atomic_read 1577 fs/gfs2/rgrp.c extlen = max_t(u32, atomic_read(&ip->i_sizehint), ap->target); atomic_read 621 fs/gfs2/super.c wait_event(sdp->sd_reserving_log_wait, atomic_read(&sdp->sd_reserving_log) == 0); atomic_read 622 fs/gfs2/super.c gfs2_assert_warn(sdp, atomic_read(&sdp->sd_log_blks_free) == sdp->sd_jdesc->jd_blocks); atomic_read 761 fs/gfs2/super.c if (atomic_read(&sdp->sd_freeze_state) != SFS_UNFROZEN) atomic_read 799 fs/gfs2/super.c if (atomic_read(&sdp->sd_freeze_state) != SFS_FROZEN || atomic_read 1148 fs/gfs2/super.c if (atomic_read(&gl->gl_revokes) == 0) { atomic_read 194 fs/gfs2/trans.c enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); atomic_read 302 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); atomic_read 449 fs/hfs/bnode.c atomic_read(&node->refcnt)); atomic_read 462 fs/hfs/bnode.c atomic_read(&node->refcnt)); atomic_read 463 fs/hfs/bnode.c BUG_ON(!atomic_read(&node->refcnt)); atomic_read 148 fs/hfs/btree.c if (atomic_read(&node->refcnt)) atomic_read 151 fs/hfs/btree.c atomic_read(&node->refcnt)); atomic_read 102 fs/hfs/inode.c else if (atomic_read(&node->refcnt)) atomic_read 117 fs/hfs/inode.c if (atomic_read(&node->refcnt)) { atomic_read 468 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); atomic_read 617 fs/hfsplus/bnode.c atomic_read(&node->refcnt)); atomic_read 630 fs/hfsplus/bnode.c atomic_read(&node->refcnt)); atomic_read 631 fs/hfsplus/bnode.c BUG_ON(!atomic_read(&node->refcnt)); atomic_read 269 fs/hfsplus/btree.c if (atomic_read(&node->refcnt)) atomic_read 273 fs/hfsplus/btree.c atomic_read(&node->refcnt)); atomic_read 373 fs/hfsplus/dir.c atomic_read(&HFSPLUS_I(inode)->opencnt)) { atomic_read 396 fs/hfsplus/dir.c if (!atomic_read(&HFSPLUS_I(inode)->opencnt)) { atomic_read 98 fs/hfsplus/inode.c else if (atomic_read(&node->refcnt)) atomic_read 114 fs/hfsplus/inode.c if (atomic_read(&node->refcnt)) { atomic_read 138 fs/hfsplus/xattr.c switch (atomic_read(&sbi->attr_tree_state)) { atomic_read 187 fs/hpfs/inode.c if (hpfs_inode->i_rddir_off && !atomic_read(&i->i_count)) { atomic_read 442 fs/inode.c !atomic_read(&inode->i_count) && inode->i_sb->s_flags & SB_ACTIVE) atomic_read 632 fs/inode.c if (atomic_read(&inode->i_count)) atomic_read 693 fs/inode.c if (atomic_read(&inode->i_count)) { atomic_read 749 fs/inode.c if (atomic_read(&inode->i_count) || atomic_read 2103 fs/inode.c if (atomic_read(&inode->i_dio_count)) atomic_read 2105 fs/inode.c } while (atomic_read(&inode->i_dio_count)); atomic_read 2121 fs/inode.c if (atomic_read(&inode->i_dio_count)) atomic_read 435 fs/io_uring.c + atomic_read(&ctx->cached_cq_overflow); atomic_read 2340 fs/io_uring.c if (!atomic_read(&list->cnt)) atomic_read 2350 fs/io_uring.c if (!atomic_read(&list->cnt)) { atomic_read 2951 fs/io_uring.c atomic_read(&ctx->cq_timeouts) != iowq->nr_timeouts; atomic_read 3002 fs/io_uring.c iowq.nr_timeouts = atomic_read(&ctx->cq_timeouts); atomic_read 52 fs/iomap/buffered-io.c WARN_ON_ONCE(atomic_read(&iop->read_count)); atomic_read 53 fs/iomap/buffered-io.c WARN_ON_ONCE(atomic_read(&iop->write_count)); atomic_read 671 fs/jbd2/checkpoint.c J_ASSERT(atomic_read(&transaction->t_updates) == 0); atomic_read 69 fs/jbd2/commit.c if (atomic_read(&bh->b_count) != 1) atomic_read 438 fs/jbd2/commit.c while (atomic_read(&commit_transaction->t_updates)) { atomic_read 443 fs/jbd2/commit.c if (atomic_read(&commit_transaction->t_updates)) { atomic_read 456 fs/jbd2/commit.c J_ASSERT (atomic_read(&commit_transaction->t_outstanding_credits) <= atomic_read 518 fs/jbd2/commit.c atomic_sub(atomic_read(&journal->j_reserved_credits), atomic_read 564 fs/jbd2/commit.c atomic_read(&commit_transaction->t_outstanding_credits); atomic_read 568 fs/jbd2/commit.c atomic_read(&commit_transaction->t_outstanding_credits)); atomic_read 824 fs/jbd2/commit.c J_ASSERT_BH(bh, atomic_read(&bh->b_count) == 0); atomic_read 1093 fs/jbd2/commit.c atomic_read(&commit_transaction->t_handle_count); atomic_read 2477 fs/jbd2/journal.c (atomic_read(&bh->b_count) > 0) || atomic_read 2717 fs/jbd2/journal.c int n = atomic_read(&nr_journal_heads); atomic_read 91 fs/jbd2/transaction.c atomic_read(&journal->j_reserved_credits)); atomic_read 236 fs/jbd2/transaction.c if (atomic_read(&journal->j_reserved_credits) + total > atomic_read 241 fs/jbd2/transaction.c atomic_read(&journal->j_reserved_credits) + total <= atomic_read 284 fs/jbd2/transaction.c atomic_read(&journal->j_reserved_credits) + rsv_blocks atomic_read 413 fs/jbd2/transaction.c atomic_read(&transaction->t_outstanding_credits), atomic_read 692 fs/jbd2/transaction.c J_ASSERT(atomic_read(&transaction->t_updates) > 0); atomic_read 756 fs/jbd2/transaction.c if (atomic_read(&journal->j_reserved_credits)) { atomic_read 759 fs/jbd2/transaction.c atomic_read(&journal->j_reserved_credits) == 0); atomic_read 773 fs/jbd2/transaction.c if (!atomic_read(&transaction->t_updates)) { atomic_read 1736 fs/jbd2/transaction.c J_ASSERT(atomic_read(&transaction->t_updates) > 0); atomic_read 1822 fs/jbd2/transaction.c (atomic_read(&transaction->t_outstanding_credits) > atomic_read 865 fs/jffs2/xattr.c if (!atomic_read(&xd->refcnt)) { atomic_read 1325 fs/jffs2/xattr.c if (atomic_read(&xd->refcnt) || xd->node != (void *)xd) atomic_read 609 fs/jfs/jfs_dmap.c if ((atomic_read(&bmp->db_active[agpref]) == 0) && atomic_read 620 fs/jfs/jfs_dmap.c if (atomic_read(&bmp->db_active[agpref])) atomic_read 755 fs/jfs/jfs_dmap.c if (atomic_read(&bmp->db_active[agno])) atomic_read 789 fs/jfs/jfs_dmap.c writers = atomic_read(&bmp->db_active[agno]); atomic_read 229 fs/jfs/jfs_imap.c dinom_le->in_numinos = cpu_to_le32(atomic_read(&imp->im_numinos)); atomic_read 230 fs/jfs/jfs_imap.c dinom_le->in_numfree = cpu_to_le32(atomic_read(&imp->im_numfree)); atomic_read 1361 fs/jfs/jfs_imap.c if (atomic_read(&JFS_SBI(pip->i_sb)->bmap->db_active[agno])) { atomic_read 2851 fs/jfs/jfs_imap.c imap->im_nextiag, atomic_read(&imap->im_numinos), atomic_read 2852 fs/jfs/jfs_imap.c atomic_read(&imap->im_numfree)); atomic_read 2957 fs/jfs/jfs_imap.c if (xnuminos != atomic_read(&imap->im_numinos) || atomic_read 2958 fs/jfs/jfs_imap.c xnumfree != atomic_read(&imap->im_numfree)) { atomic_read 140 fs/jfs/super.c maxinodes = min((s64) atomic_read(&imap->im_numinos) + atomic_read 144 fs/jfs/super.c buf->f_ffree = maxinodes - (atomic_read(&imap->im_numinos) - atomic_read 145 fs/jfs/super.c atomic_read(&imap->im_numfree)); atomic_read 30 fs/kernfs/dir.c return atomic_read(&kn->active) >= 0; atomic_read 469 fs/kernfs/dir.c if (atomic_read(&kn->active) != KN_DEACTIVATED_BIAS) atomic_read 475 fs/kernfs/dir.c atomic_read(&kn->active) == KN_DEACTIVATED_BIAS); atomic_read 494 fs/kernfs/dir.c WARN_ON(!atomic_read(&kn->count)); atomic_read 525 fs/kernfs/dir.c WARN_ONCE(atomic_read(&kn->active) != KN_DEACTIVATED_BIAS, atomic_read 527 fs/kernfs/dir.c parent ? parent->name : "", kn->name, atomic_read(&kn->active)); atomic_read 1269 fs/kernfs/dir.c WARN_ON_ONCE(atomic_read(&pos->active) != KN_DEACTIVATED_BIAS); atomic_read 1321 fs/kernfs/dir.c WARN_ON_ONCE(atomic_read(&kn->active) != KN_DEACTIVATED_BIAS); atomic_read 1466 fs/kernfs/dir.c atomic_read(&kn->active) == KN_DEACTIVATED_BIAS) atomic_read 165 fs/kernfs/file.c of->event = atomic_read(&of->kn->attr.open->event); atomic_read 210 fs/kernfs/file.c of->event = atomic_read(&of->kn->attr.open->event); atomic_read 841 fs/kernfs/file.c if (of->event != atomic_read(&on->event)) atomic_read 366 fs/lockd/svc.c wait_event(nlm_ntf_wq, atomic_read(&nlm_ntf_refcnt) == 0); atomic_read 1824 fs/locks.c if (atomic_read(&inode->i_writecount) != self_wcount || atomic_read 1825 fs/locks.c atomic_read(&inode->i_readcount) != self_rcount) atomic_read 409 fs/mbcache.c WARN_ON(atomic_read(&entry->e_refcnt) != 1); atomic_read 455 fs/nfs/file.c if (atomic_read(&nfsi->commit_info.rpcs_out)) { atomic_read 571 fs/nfs/inode.c atomic_read(&inode->i_count)); atomic_read 1809 fs/nfs/inode.c atomic_read(&inode->i_count), fattr->valid); atomic_read 103 fs/nfs/pagelist.c !atomic_read(&l_ctx->io_count)); atomic_read 121 fs/nfs/pagelist.c if (atomic_read(&l_ctx->io_count) > 0) { atomic_read 126 fs/nfs/pagelist.c if (atomic_read(&l_ctx->io_count) == 0) { atomic_read 527 fs/nfs/pnfs.c if (atomic_read(&lo->plh_outstanding) == 0) atomic_read 929 fs/nfs/pnfs.c new_barrier = newseq - atomic_read(&lo->plh_outstanding); atomic_read 1128 fs/nfs/pnfs.c if (atomic_read(&lo->plh_outstanding) != 0) atomic_read 1950 fs/nfs/pnfs.c atomic_read(&lo->plh_outstanding) != 0) { atomic_read 1953 fs/nfs/pnfs.c !atomic_read(&lo->plh_outstanding))); atomic_read 86 fs/nfs/pnfs_dev.c if (atomic_read(&d->ref)) atomic_read 332 fs/nfs/pnfs_dev.c if (d->nfs_client == clp && atomic_read(&d->ref)) { atomic_read 1668 fs/nfs/write.c !atomic_read(&cinfo->rpcs_out)); atomic_read 1993 fs/nfs/write.c if (!atomic_read(&nfsi->commit_info.rpcs_out)) atomic_read 378 fs/nfsd/filecache.c if (atomic_read(&nf->nf_ref) > 1) atomic_read 1660 fs/nfsd/nfs4proc.c if (atomic_read(&ls->ls_stid.sc_file->fi_lo_recalls)) atomic_read 131 fs/nfsd/nfs4state.c if (atomic_read(&ses->se_ref) > ref_held_by_me) atomic_read 582 fs/nfsd/nfs4state.c atomic_read(&fp->fi_access[O_RDONLY])) atomic_read 586 fs/nfsd/nfs4state.c atomic_read(&fp->fi_access[O_WRONLY])) atomic_read 601 fs/nfsd/nfs4state.c if (atomic_read(&fp->fi_access[1 - oflag]) == 0) atomic_read 1946 fs/nfsd/nfs4state.c WARN_ON_ONCE(atomic_read(&ses->se_ref)); atomic_read 1996 fs/nfsd/nfs4state.c if (atomic_read(&clp->cl_rpc_users)) atomic_read 2552 fs/nfsd/nfs4state.c wait_event(expiry_wq, atomic_read(&clp->cl_rpc_users) == 0); atomic_read 5194 fs/nfsd/nfs4state.c atomic_read(&nn->nr_reclaim_complete) == atomic_read 240 fs/nfsd/nfscache.c if (atomic_read(&nn->num_drc_entries) <= nn->max_drc_entries && atomic_read 277 fs/nfsd/nfscache.c return atomic_read(&nn->num_drc_entries); atomic_read 368 fs/nfsd/nfscache.c nn->longest_chain_cachesize = atomic_read(&nn->num_drc_entries); atomic_read 373 fs/nfsd/nfscache.c atomic_read(&nn->num_drc_entries)); atomic_read 578 fs/nfsd/nfscache.c atomic_read(&nn->num_drc_entries)); atomic_read 506 fs/nfsd/nfssvc.c wait_event(nn->ntf_wq, atomic_read(&nn->ntf_refcnt) == 0); atomic_read 198 fs/nfsd/trace.h __entry->nf_ref = atomic_read(&nf->nf_ref); atomic_read 247 fs/nfsd/trace.h __entry->nf_ref = nf ? atomic_read(&nf->nf_ref) : 0; atomic_read 933 fs/nfsd/vfs.c if (atomic_read(&inode->i_writecount) > 1 atomic_read 180 fs/nilfs2/page.c i++, bh, atomic_read(&bh->b_count), atomic_read 350 fs/nilfs2/segbuf.c if (unlikely(atomic_read(&segbuf->sb_err))) { atomic_read 507 fs/nilfs2/segbuf.c if (unlikely(atomic_read(&segbuf->sb_err) > 0)) { atomic_read 280 fs/nilfs2/segment.c if (atomic_read(&nilfs->ns_ndirtyblks) > sci->sc_watermark) atomic_read 1401 fs/nilfs2/segment.c if (atomic_read(&segbuf->sb_err)) { atomic_read 1420 fs/nilfs2/segment.c if (atomic_read(&segbuf->sb_err) && atomic_read 2177 fs/nilfs2/segment.c if (atomic_read(&wait_req.done)) { atomic_read 2199 fs/nilfs2/segment.c if (!atomic_read(&wrq->done) && atomic_read 2204 fs/nilfs2/segment.c if (atomic_read(&wrq->done)) { atomic_read 632 fs/nilfs2/sysfs.c ndirtyblks = atomic_read(&nilfs->ns_ndirtyblks); atomic_read 710 fs/nilfs2/the_nilfs.c nincsegs = atomic_read(&nilfs->ns_ndirtyblks) / atomic_read 685 fs/notify/fanotify/fanotify_user.c if (atomic_read(&group->num_marks) > group->fanotify_data.max_marks) atomic_read 803 fs/notify/fanotify/fanotify_user.c if (atomic_read(&user->fanotify_listeners) > FANOTIFY_DEFAULT_MAX_LISTENERS) { atomic_read 64 fs/notify/fsnotify.c if (!atomic_read(&inode->i_count)) { atomic_read 67 fs/notify/group.c wait_event(group->notification_waitq, !atomic_read(&group->user_waits)); atomic_read 2798 fs/ntfs/mft.c if (atomic_read(&ni->count) > 2) { atomic_read 6095 fs/ocfs2/alloc.c atomic_read(&osb->osb_tl_disable) == 0) { atomic_read 325 fs/ocfs2/cluster/heartbeat.c if (atomic_read(®->hr_steady_iterations) != 0) atomic_read 855 fs/ocfs2/cluster/heartbeat.c if (atomic_read(®->hr_steady_iterations) != 0) atomic_read 1169 fs/ocfs2/cluster/heartbeat.c if (atomic_read(®->hr_steady_iterations) != 0) { atomic_read 1176 fs/ocfs2/cluster/heartbeat.c if (atomic_read(®->hr_steady_iterations) != 0) { atomic_read 1880 fs/ocfs2/cluster/heartbeat.c atomic_read(®->hr_steady_iterations) == 0 || atomic_read 2101 fs/ocfs2/cluster/heartbeat.c ((atomic_read(®->hr_steady_iterations) == 0) ? atomic_read 2110 fs/ocfs2/cluster/heartbeat.c if (atomic_read(®->hr_steady_iterations) != 0) { atomic_read 485 fs/ocfs2/cluster/tcp.c return atomic_read(&o2net_connected_peers); atomic_read 1547 fs/ocfs2/cluster/tcp.c if (atomic_read(&nn->nn_timeout)) { atomic_read 1603 fs/ocfs2/cluster/tcp.c timeout = atomic_read(&nn->nn_timeout); atomic_read 1748 fs/ocfs2/cluster/tcp.c BUG_ON(atomic_read(&o2net_connected_peers) < 0); atomic_read 102 fs/ocfs2/dlm/dlmdebug.c res->inflight_locks, atomic_read(&res->asts_reserved)); atomic_read 508 fs/ocfs2/dlm/dlmdebug.c atomic_read(&res->asts_reserved), atomic_read 722 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->res_cur_count), atomic_read 723 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->res_tot_count)); atomic_read 726 fs/ocfs2/dlm/dlmdebug.c tot_mles += atomic_read(&dlm->mle_tot_count[i]); atomic_read 729 fs/ocfs2/dlm/dlmdebug.c cur_mles += atomic_read(&dlm->mle_cur_count[i]); atomic_read 738 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->mle_cur_count[DLM_MLE_BLOCK]), atomic_read 739 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->mle_tot_count[DLM_MLE_BLOCK])); atomic_read 744 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->mle_cur_count[DLM_MLE_MASTER]), atomic_read 745 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->mle_tot_count[DLM_MLE_MASTER])); atomic_read 750 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->mle_cur_count[DLM_MLE_MIGRATION]), atomic_read 751 fs/ocfs2/dlm/dlmdebug.c atomic_read(&dlm->mle_tot_count[DLM_MLE_MIGRATION])); atomic_read 1107 fs/ocfs2/dlm/dlmmaster.c (atomic_read(&mle->woken) == 1), atomic_read 2686 fs/ocfs2/dlm/dlmmaster.c (atomic_read(&mle->woken) == 1), atomic_read 2690 fs/ocfs2/dlm/dlmmaster.c if (atomic_read(&mle->woken) == 1 || atomic_read 305 fs/ocfs2/journal.c flushed = atomic_read(&journal->j_num_trans); atomic_read 323 fs/ocfs2/journal.c flushed = atomic_read(&journal->j_num_trans); atomic_read 978 fs/ocfs2/journal.c num_running_trans = atomic_read(&(osb->journal->j_num_trans)); atomic_read 997 fs/ocfs2/journal.c BUG_ON(atomic_read(&(osb->journal->j_num_trans)) != 0); atomic_read 1929 fs/ocfs2/journal.c if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE) atomic_read 1933 fs/ocfs2/journal.c atomic_read(&os->os_state)); atomic_read 1943 fs/ocfs2/journal.c if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE) atomic_read 1965 fs/ocfs2/journal.c atomic_read(&os->os_state)); atomic_read 1981 fs/ocfs2/journal.c if (atomic_read(&os->os_state) == ORPHAN_SCAN_ACTIVE) atomic_read 1992 fs/ocfs2/journal.c if (atomic_read(&os->os_state) == ORPHAN_SCAN_ACTIVE) { atomic_read 2274 fs/ocfs2/journal.c (!quota && atomic_read(&osb->vol_state) == VOLUME_MOUNTED) || atomic_read 2275 fs/ocfs2/journal.c atomic_read(&osb->vol_state) == VOLUME_MOUNTED_QUOTAS || atomic_read 2276 fs/ocfs2/journal.c atomic_read(&osb->vol_state) == VOLUME_DISABLED); atomic_read 2281 fs/ocfs2/journal.c if (atomic_read(&osb->vol_state) == VOLUME_DISABLED) { atomic_read 2301 fs/ocfs2/journal.c atomic_read(&journal->j_num_trans) == 0)) { atomic_read 2304 fs/ocfs2/journal.c atomic_read(&journal->j_num_trans) atomic_read 2323 fs/ocfs2/journal.c if (kthread_should_stop() && atomic_read(&journal->j_num_trans)){ atomic_read 2327 fs/ocfs2/journal.c atomic_read(&journal->j_num_trans)); atomic_read 215 fs/ocfs2/stack_user.c if ((c->oc_type == NO_CONTROLD) || atomic_read(&ocfs2_control_opened)) atomic_read 1037 fs/ocfs2/stack_user.c wait_event(lc->oc_wait, (atomic_read(&lc->oc_this_node) > 0)); atomic_read 1071 fs/ocfs2/stack_user.c rc = atomic_read(&lc->oc_this_node); atomic_read 976 fs/ocfs2/suballoc.c atomic_read(&osb->s_num_meta_stolen) < OCFS2_MAX_TO_STEAL) atomic_read 1065 fs/ocfs2/suballoc.c atomic_read(&osb->s_num_inodes_stolen) < OCFS2_MAX_TO_STEAL) atomic_read 230 fs/ocfs2/super.c atomic_read(&osb->vol_state), osb->osb_flags); atomic_read 290 fs/ocfs2/super.c atomic_read(&osb->journal->j_num_trans)); atomic_read 296 fs/ocfs2/super.c atomic_read(&osb->alloc_stats.bitmap_data), atomic_read 297 fs/ocfs2/super.c atomic_read(&osb->alloc_stats.local_data), atomic_read 298 fs/ocfs2/super.c atomic_read(&osb->alloc_stats.bg_allocs), atomic_read 299 fs/ocfs2/super.c atomic_read(&osb->alloc_stats.moves), atomic_read 300 fs/ocfs2/super.c atomic_read(&osb->alloc_stats.bg_extends)); atomic_read 314 fs/ocfs2/super.c atomic_read(&osb->s_num_inodes_stolen), atomic_read 316 fs/ocfs2/super.c atomic_read(&osb->s_num_meta_stolen)); atomic_read 323 fs/ocfs2/super.c if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE) atomic_read 35 fs/overlayfs/inode.c if (atomic_read(&realinode->i_writecount) < 0) atomic_read 287 fs/proc/array.c qsize = atomic_read(&__task_cred(p)->user->sigpending); atomic_read 1079 fs/proc/base.c if (atomic_read(&p->mm->mm_users) > 1) { atomic_read 228 fs/proc/generic.c if (atomic_read(&PDE(d_inode(dentry))->in_use) < 0) atomic_read 235 fs/proc/generic.c return atomic_read(&PDE(d_inode(dentry))->in_use) < 0; atomic_read 677 fs/proc/proc_sysctl.c if (event != atomic_read(&table->poll->event)) { atomic_read 42 fs/proc/task_nommu.c if (atomic_read(&mm->mm_count) > 1 || atomic_read 52 fs/proc/task_nommu.c if (atomic_read(&mm->mm_count) > 1) atomic_read 62 fs/proc/task_nommu.c if (current->files && atomic_read(¤t->files->count) > 1) atomic_read 44 fs/pstore/ram_core.c return atomic_read(&prz->buffer->size); atomic_read 49 fs/pstore/ram_core.c return atomic_read(&prz->buffer->start); atomic_read 62 fs/pstore/ram_core.c old = atomic_read(&prz->buffer->start); atomic_read 84 fs/pstore/ram_core.c old = atomic_read(&prz->buffer->size); atomic_read 548 fs/quota/dquot.c if (atomic_read(&dquot->dq_count)) { atomic_read 560 fs/quota/dquot.c atomic_read(&dquot->dq_count) == 1); atomic_read 761 fs/quota/dquot.c if (!atomic_read(&dquot->dq_count)) { atomic_read 771 fs/quota/dquot.c if (atomic_read(&dquot->dq_count) > 1) { atomic_read 776 fs/quota/dquot.c atomic_read(&dquot->dq_count) == 1) atomic_read 888 fs/quota/dquot.c if (!atomic_read(&dquot->dq_count)) atomic_read 960 fs/quota/dquot.c !atomic_read(&inode->i_writecount) || atomic_read 1445 fs/reiserfs/bitmap.c BUG_ON(atomic_read(&bh->b_count) == 0); atomic_read 2318 fs/reiserfs/fix_node.c if (atomic_read(&(bh->b_count)) <= 0) atomic_read 741 fs/reiserfs/journal.c if (atomic_read(&nr_reiserfs_jh) <= 0) atomic_read 904 fs/reiserfs/journal.c atomic_read(&other_jl->j_older_commits_done)) atomic_read 922 fs/reiserfs/journal.c if (atomic_read(&other_jl->j_commit_left) != 0) { atomic_read 954 fs/reiserfs/journal.c if (atomic_read(&j->j_async_throttle)) { atomic_read 986 fs/reiserfs/journal.c if (atomic_read(&jl->j_older_commits_done)) { atomic_read 1018 fs/reiserfs/journal.c if (atomic_read(&jl->j_commit_left) <= 0) { atomic_read 1099 fs/reiserfs/journal.c BUG_ON(atomic_read(&jl->j_commit_left) != 1); atomic_read 1363 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) != 0) { atomic_read 1365 fs/reiserfs/journal.c atomic_read(&journal->j_wcount)); atomic_read 1383 fs/reiserfs/journal.c if (atomic_read(&jl->j_nonzerolen) <= 0 && atomic_read 1384 fs/reiserfs/journal.c atomic_read(&jl->j_commit_left) <= 0) { atomic_read 1399 fs/reiserfs/journal.c if (atomic_read(&jl->j_nonzerolen) <= 0 && atomic_read 1400 fs/reiserfs/journal.c atomic_read(&jl->j_commit_left) <= 0) { atomic_read 1408 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) != 0) { atomic_read 1462 fs/reiserfs/journal.c if (atomic_read(&pjl->j_commit_left)) atomic_read 1516 fs/reiserfs/journal.c if (atomic_read(&saved_bh->b_count) < 0) { atomic_read 1638 fs/reiserfs/journal.c if (jl->j_len == 0 || atomic_read(&jl->j_nonzerolen) == 0) { atomic_read 1744 fs/reiserfs/journal.c atomic_read(&jl->j_commit_left) atomic_read 1805 fs/reiserfs/journal.c if (atomic_read(&tjl->j_commit_left) || atomic_read 1809 fs/reiserfs/journal.c cur_len = atomic_read(&tjl->j_nonzerolen); atomic_read 2915 fs/reiserfs/journal.c atomic_read(&journal->j_jlock) || atomic_read 2993 fs/reiserfs/journal.c while ((atomic_read(&journal->j_wcount) > 0 || atomic_read 2994 fs/reiserfs/journal.c atomic_read(&journal->j_jlock)) && atomic_read 3062 fs/reiserfs/journal.c || (!join && atomic_read(&journal->j_wcount) > 0 atomic_read 3066 fs/reiserfs/journal.c && atomic_read(&journal->j_jlock)) atomic_read 3077 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) > 10) { atomic_read 3087 fs/reiserfs/journal.c if (atomic_read(&journal->j_jlock)) { atomic_read 3089 fs/reiserfs/journal.c atomic_read(&journal->j_jlock)) { atomic_read 3308 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) <= 0) { atomic_read 3311 fs/reiserfs/journal.c atomic_read(&journal->j_wcount)); atomic_read 3449 fs/reiserfs/journal.c if (atomic_read(&bh->b_count) < 0) { atomic_read 3498 fs/reiserfs/journal.c atomic_read(&cur->jlist->j_commit_left) > 0 && cur->bh && atomic_read 3568 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) <= 0 && atomic_read 3624 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) > 0) atomic_read 3643 fs/reiserfs/journal.c if (atomic_read(&journal->j_wcount) > 0) { atomic_read 3662 fs/reiserfs/journal.c if (atomic_read(&journal->j_jlock)) { atomic_read 3694 fs/reiserfs/journal.c if (!(journal->j_must_wait > 0) && !(atomic_read(&journal->j_jlock)) atomic_read 3792 fs/reiserfs/journal.c if (atomic_read atomic_read 3883 fs/reiserfs/journal.c if (atomic_read(&jl->j_commit_left) > 1) atomic_read 161 fs/reiserfs/prints.c atomic_read(&(bh->b_count)), atomic_read 673 fs/reiserfs/prints.c (tbSh) ? atomic_read(&tbSh->b_count) : -1, atomic_read 675 fs/reiserfs/prints.c (tb->L[h]) ? atomic_read(&tb->L[h]->b_count) : -1, atomic_read 677 fs/reiserfs/prints.c (tb->R[h]) ? atomic_read(&tb->R[h]->b_count) : -1, atomic_read 719 fs/reiserfs/prints.c tb->FEB[i] ? atomic_read(&tb->FEB[i]->b_count) : 0, atomic_read 117 fs/reiserfs/procfs.c atomic_read(&r->s_generation_counter), atomic_read 372 fs/reiserfs/procfs.c atomic_read(&r->s_journal->j_wcount), atomic_read 2305 fs/reiserfs/reiserfs.h #define get_generation(s) atomic_read (&fs_generation(s)) atomic_read 922 fs/reiserfs/stree.c #define held_by_others(bh) (atomic_read(&(bh)->b_count) > 1) atomic_read 1554 fs/reiserfs/stree.c if (atomic_read(&inode->i_count) > 1 || atomic_read 346 fs/ubifs/super.c ubifs_assert(c, !atomic_read(&inode->i_count)); atomic_read 230 fs/udf/file.c atomic_read(&inode->i_writecount) == 1) { atomic_read 48 fs/xfs/libxfs/xfs_sb.c ASSERT(atomic_read(&pag->pag_ref) >= 0); atomic_read 88 fs/xfs/libxfs/xfs_sb.c ASSERT(atomic_read(&pag->pag_ref) > 0); atomic_read 72 fs/xfs/xfs_aops.c ASSERT(!iop || atomic_read(&iop->write_count) > 0); atomic_read 881 fs/xfs/xfs_aops.c ASSERT(!iop || atomic_read(&iop->write_count) == 0); atomic_read 52 fs/xfs/xfs_bmap_item.c ASSERT(atomic_read(&buip->bui_refcount) > 0); atomic_read 87 fs/xfs/xfs_bmap_item.c ASSERT(atomic_read(&buip->bui_next_extent) == atomic_read 164 fs/xfs/xfs_buf.c ASSERT(atomic_read(&bp->b_hold) >= 1); atomic_read 993 fs/xfs/xfs_buf.c ASSERT(atomic_read(&bp->b_hold) > 0); atomic_read 1014 fs/xfs/xfs_buf.c if ((atomic_read(&bp->b_hold) == 1) && !list_empty(&bp->b_lru)) atomic_read 1021 fs/xfs/xfs_buf.c if (!(bp->b_flags & XBF_STALE) && atomic_read(&bp->b_lru_ref)) { atomic_read 1101 fs/xfs/xfs_buf.c if (atomic_read(&bp->b_pin_count) && (bp->b_flags & XBF_STALE)) atomic_read 1124 fs/xfs/xfs_buf.c if (atomic_read(&bp->b_pin_count) == 0) atomic_read 1130 fs/xfs/xfs_buf.c if (atomic_read(&bp->b_pin_count) == 0) atomic_read 1567 fs/xfs/xfs_buf.c if (atomic_read(&bp->b_hold) > 1) { atomic_read 311 fs/xfs/xfs_buf.h if (!list_empty(&bp->b_lru) || atomic_read(&bp->b_lru_ref) > 1) atomic_read 318 fs/xfs/xfs_buf.h return atomic_read(&bp->b_pin_count); atomic_read 127 fs/xfs/xfs_buf_item.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 306 fs/xfs/xfs_buf_item.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 365 fs/xfs/xfs_buf_item.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 401 fs/xfs/xfs_buf_item.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 92 fs/xfs/xfs_dquot_item.c ASSERT(atomic_read(&dqp->q_pincount) > 0); atomic_read 106 fs/xfs/xfs_dquot_item.c if (atomic_read(&dqp->q_pincount) == 0) atomic_read 113 fs/xfs/xfs_dquot_item.c wait_event(dqp->q_pinwait, (atomic_read(&dqp->q_pincount) == 0)); atomic_read 145 fs/xfs/xfs_dquot_item.c if (atomic_read(&dqp->q_pincount) > 0) atomic_read 170 fs/xfs/xfs_dquot_item.c if (atomic_read(&dqp->q_pincount) > 0) { atomic_read 56 fs/xfs/xfs_extfree_item.c ASSERT(atomic_read(&efip->efi_refcount) > 0); atomic_read 101 fs/xfs/xfs_extfree_item.c ASSERT(atomic_read(&efip->efi_next_extent) == atomic_read 756 fs/xfs/xfs_file.c atomic_read(&page->_refcount) == 1, TASK_INTERRUPTIBLE, atomic_read 78 fs/xfs/xfs_filestream.c ret = atomic_read(&pag->pagf_fstrms); atomic_read 55 fs/xfs/xfs_icache.c ASSERT(atomic_read(&ip->i_pincount) == 0); atomic_read 115 fs/xfs/xfs_icache.c ASSERT(atomic_read(&ip->i_pincount) == 0); atomic_read 1734 fs/xfs/xfs_icache.c atomic_read(&VFS_I(ip)->i_dio_count)) atomic_read 1522 fs/xfs/xfs_inode.c ASSERT(!atomic_read(&VFS_I(ip)->i_count) || atomic_read 445 fs/xfs/xfs_inode.h #define xfs_ipincount(ip) ((unsigned int) atomic_read(&ip->i_pincount)) atomic_read 463 fs/xfs/xfs_inode_item.c ASSERT(atomic_read(&ip->i_pincount) > 0); atomic_read 1870 fs/xfs/xfs_log.c ASSERT(atomic_read(&iclog->ic_refcnt) == 0); atomic_read 2936 fs/xfs/xfs_log.c ASSERT(atomic_read(&iclog->ic_refcnt) == 0); atomic_read 3184 fs/xfs/xfs_log.c ASSERT(atomic_read(&iclog->ic_refcnt) > 0); atomic_read 3317 fs/xfs/xfs_log.c atomic_read(&iclog->ic_refcnt) == 0 && iclog->ic_offset == 0)) { atomic_read 3331 fs/xfs/xfs_log.c if (atomic_read(&iclog->ic_refcnt) == 0) { atomic_read 3542 fs/xfs/xfs_log.c ASSERT(atomic_read(&ticket->t_ref) > 0); atomic_read 3551 fs/xfs/xfs_log.c ASSERT(atomic_read(&ticket->t_ref) > 0); atomic_read 129 fs/xfs/xfs_mount.c ASSERT(atomic_read(&pag->pag_ref) == 0); atomic_read 148 fs/xfs/xfs_mount.c ASSERT(atomic_read(&pag->pag_ref) == 0); atomic_read 117 fs/xfs/xfs_pwork.c atomic_read(&pctl->nr_work) == 0, HZ) == 0) atomic_read 155 fs/xfs/xfs_qm.c ASSERT(atomic_read(&dqp->q_pincount) == 0); atomic_read 51 fs/xfs/xfs_refcount_item.c ASSERT(atomic_read(&cuip->cui_refcount) > 0); atomic_read 86 fs/xfs/xfs_refcount_item.c ASSERT(atomic_read(&cuip->cui_next_extent) == atomic_read 51 fs/xfs/xfs_rmap_item.c ASSERT(atomic_read(&ruip->rui_refcount) > 0); atomic_read 85 fs/xfs/xfs_rmap_item.c ASSERT(atomic_read(&ruip->rui_next_extent) == atomic_read 1184 fs/xfs/xfs_super.c while (atomic_read(&mp->m_active_trans) > 0) atomic_read 1203 fs/xfs/xfs_super.c WARN_ON(atomic_read(&mp->m_active_trans) != 0); atomic_read 293 fs/xfs/xfs_trace.h __entry->hold = atomic_read(&bp->b_hold); atomic_read 294 fs/xfs/xfs_trace.h __entry->pincount = atomic_read(&bp->b_pin_count); atomic_read 362 fs/xfs/xfs_trace.h __entry->hold = atomic_read(&bp->b_hold); atomic_read 363 fs/xfs/xfs_trace.h __entry->pincount = atomic_read(&bp->b_pin_count); atomic_read 405 fs/xfs/xfs_trace.h __entry->hold = atomic_read(&bp->b_hold); atomic_read 406 fs/xfs/xfs_trace.h __entry->pincount = atomic_read(&bp->b_pin_count); atomic_read 445 fs/xfs/xfs_trace.h __entry->bli_refcount = atomic_read(&bip->bli_refcount); atomic_read 449 fs/xfs/xfs_trace.h __entry->buf_hold = atomic_read(&bip->bli_buf->b_hold); atomic_read 450 fs/xfs/xfs_trace.h __entry->buf_pincount = atomic_read(&bip->bli_buf->b_pin_count); atomic_read 697 fs/xfs/xfs_trace.h __entry->count = atomic_read(&VFS_I(ip)->i_count); atomic_read 698 fs/xfs/xfs_trace.h __entry->pincount = atomic_read(&ip->i_pincount); atomic_read 146 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 197 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 294 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 391 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 443 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 464 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 498 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 587 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 641 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 665 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 690 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 712 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 745 fs/xfs/xfs_trans_buf.c ASSERT(atomic_read(&bip->bli_refcount) > 0); atomic_read 29 include/asm-generic/atomic-instrumented.h #define atomic_read atomic_read atomic_read 522 include/asm-generic/atomic-long.h return atomic_read(v); atomic_read 170 include/asm-generic/atomic.h #ifndef atomic_read atomic_read 42 include/asm-generic/qrwlock.h cnts = atomic_read(&lock->cnts); atomic_read 61 include/asm-generic/qrwlock.h cnts = atomic_read(&lock->cnts); atomic_read 26 include/asm-generic/qspinlock.h return atomic_read(&lock->val); atomic_read 41 include/asm-generic/qspinlock.h return !atomic_read(&lock.val); atomic_read 51 include/asm-generic/qspinlock.h return atomic_read(&lock->val) & ~_Q_LOCKED_MASK; atomic_read 60 include/asm-generic/qspinlock.h u32 val = atomic_read(&lock->val); atomic_read 211 include/crypto/if_alg.h atomic_read(&ctx->rcvused), 0); atomic_read 62 include/drm/spsc_queue.h return atomic_read(&queue->job_count); atomic_read 1086 include/linux/atomic-fallback.h int c = atomic_read(v); atomic_read 1136 include/linux/atomic-fallback.h int c = atomic_read(v); atomic_read 1152 include/linux/atomic-fallback.h int c = atomic_read(v); atomic_read 1168 include/linux/atomic-fallback.h int dec, c = atomic_read(v); atomic_read 310 include/linux/blk-cgroup.h if (atomic_read(&css->cgroup->congestion_count)) { atomic_read 773 include/linux/blk-cgroup.h int old = atomic_read(&blkg->use_delay); atomic_read 801 include/linux/blk-cgroup.h int old = atomic_read(&blkg->use_delay); atomic_read 661 include/linux/blkdev.h #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) atomic_read 211 include/linux/console.h WARN_ON(!atomic_read(&ignore_console_lock_warning) && \ atomic_read 112 include/linux/cpumask.h return atomic_read(&__num_online_cpus); atomic_read 216 include/linux/firewire.h return atomic_read(&device->state) == FW_DEVICE_SHUTDOWN; atomic_read 37 include/linux/freezer.h if (likely(!atomic_read(&system_freezing_cnt))) atomic_read 563 include/linux/fs.h return atomic_read(&mapping->i_mmap_writable) > 0; atomic_read 2822 include/linux/fs.h return atomic_read(&mapping->nr_thps); atomic_read 2943 include/linux/fs.h return atomic_read(&inode->i_writecount) > 0; atomic_read 2949 include/linux/fs.h BUG_ON(!atomic_read(&inode->i_readcount)); atomic_read 133 include/linux/iocontext.h WARN_ON_ONCE(atomic_read(&ioc->active_ref) <= 0); atomic_read 142 include/linux/iocontext.h WARN_ON_ONCE(atomic_read(&ioc->nr_tasks) <= 0); atomic_read 1588 include/linux/jbd2.h unsigned long committing = atomic_read(&journal-> atomic_read 254 include/linux/jump_label.h return atomic_read(&key->enabled); atomic_read 308 include/linux/jump_label.h if (atomic_read(&key->enabled) != 0) { atomic_read 309 include/linux/jump_label.h WARN_ON_ONCE(atomic_read(&key->enabled) != 1); atomic_read 319 include/linux/jump_label.h if (atomic_read(&key->enabled) != 1) { atomic_read 320 include/linux/jump_label.h WARN_ON_ONCE(atomic_read(&key->enabled) != 0); atomic_read 326 include/linux/kgdb.h (raw_smp_processor_id() == atomic_read(&kgdb_active)) atomic_read 541 include/linux/kvm_host.h int num_vcpus = atomic_read(&kvm->online_vcpus); atomic_read 551 include/linux/kvm_host.h idx < atomic_read(&kvm->online_vcpus) && \ atomic_read 1014 include/linux/kvm_host.h int slot = atomic_read(&slots->lru_slot); atomic_read 708 include/linux/mm.h return atomic_read(compound_mapcount_ptr(page)) + 1; atomic_read 735 include/linux/mm.h return atomic_read(&page->_mapcount) + 1; atomic_read 628 include/linux/mm_types.h return atomic_read(&mm->tlb_flush_pending); atomic_read 640 include/linux/mm_types.h return atomic_read(&mm->tlb_flush_pending) > 1; atomic_read 38 include/linux/osq_lock.h return atomic_read(&lock->tail) != OSQ_UNLOCKED_VAL; atomic_read 641 include/linux/page-flags.h return atomic_read(&page->_mapcount) < 0; atomic_read 645 include/linux/page-flags.h return atomic_read(&page->_mapcount) == atomic_read 646 include/linux/page-flags.h atomic_read(compound_mapcount_ptr(head)); atomic_read 67 include/linux/page_ref.h return atomic_read(&page->_refcount); atomic_read 72 include/linux/page_ref.h return atomic_read(&compound_head(page)->_refcount); atomic_read 1152 include/linux/pci.h return (atomic_read(&pdev->enable_cnt) > 0); atomic_read 52 include/linux/quotaops.h WARN_ON_ONCE(!atomic_read(&dquot->dq_count)); atomic_read 62 include/linux/quotaops.h if (atomic_read(&dquot->dq_count) > 1) atomic_read 43 include/linux/refcount.h return atomic_read(&r->refs); atomic_read 186 include/linux/rhashtable.h return atomic_read(&ht->nelems) > (tbl->size / 4 * 3) && atomic_read 199 include/linux/rhashtable.h return atomic_read(&ht->nelems) < (tbl->size * 3 / 10) && atomic_read 211 include/linux/rhashtable.h return atomic_read(&ht->nelems) > tbl->size && atomic_read 223 include/linux/rhashtable.h return atomic_read(&ht->nelems) >= ht->max_elems; atomic_read 300 include/linux/rio.h return atomic_read(&mport->state) == RIO_DEVICE_RUNNING; atomic_read 506 include/linux/sbitmap.h int old = atomic_read(index); atomic_read 522 include/linux/sbitmap.h ws = &sbq->ws[atomic_read(wait_index)]; atomic_read 369 include/linux/sched/mm.h if (likely(!(atomic_read(&mm->membarrier_state) & atomic_read 1603 include/linux/skbuff.h (atomic_read(&skb_shinfo(skb)->dataref) & SKB_DATAREF_MASK) != 1; atomic_read 1630 include/linux/skbuff.h dataref = atomic_read(&skb_shinfo(skb)->dataref); atomic_read 112 include/linux/sysctl.h return (void *)(unsigned long)atomic_read(&poll->event); atomic_read 22 include/misc/cxl-base.h return (atomic_read(&cxl_use_count) != 0); atomic_read 976 include/net/bluetooth/hci_core.h BT_DBG("hcon %p orig refcnt %d", conn, atomic_read(&conn->refcnt)); atomic_read 984 include/net/bluetooth/hci_core.h BT_DBG("hcon %p orig refcnt %d", conn, atomic_read(&conn->refcnt)); atomic_read 807 include/net/bluetooth/l2cap.h mutex_lock_nested(&chan->lock, atomic_read(&chan->nesting)); atomic_read 105 include/net/bonding.h return atomic_read(&netpoll_block_tx); atomic_read 190 include/net/dn_nsp.h return atomic_read(&sk->sk_rmem_alloc) > (sk->sk_rcvbuf >> 1); atomic_read 1266 include/net/ip_vs.h if (atomic_read(&ctl_cp->n_control) == 0) { atomic_read 1508 include/net/ip_vs.h #define IP_VS_DFWD_METHOD(dest) (atomic_read(&(dest)->conn_flags) & \ atomic_read 1682 include/net/ip_vs.h return (atomic_read(&dest->activeconns) << 8) + atomic_read 1683 include/net/ip_vs.h atomic_read(&dest->inactconns); atomic_read 221 include/net/llc_c_ev.h return atomic_read(&sk->sk_rmem_alloc) + skb->truesize < atomic_read 428 include/net/net_namespace.h return atomic_read(&net->ipv4.rt_genid); atomic_read 434 include/net/net_namespace.h return atomic_read(&net->ipv6.fib6_sernum); atomic_read 467 include/net/net_namespace.h return atomic_read(&net->fnhe_genid); atomic_read 224 include/net/request_sock.h return atomic_read(&queue->qlen); atomic_read 229 include/net/request_sock.h return atomic_read(&queue->young); atomic_read 923 include/net/sock.h unsigned int qsize = sk->sk_backlog.len + atomic_read(&sk->sk_rmem_alloc); atomic_read 1203 include/net/sock.h sk->sk_prot->name, sk, atomic_read(&sk->sk_prot->socks)); atomic_read 2051 include/net/sock.h return atomic_read(&sk->sk_rmem_alloc); atomic_read 2327 include/net/sock.h atomic_read(&sk->sk_drops) : 0; atomic_read 689 include/net/tcp.h atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf && atomic_read 1396 include/net/tcp.h atomic_read(&sk->sk_rmem_alloc)); atomic_read 1414 include/net/tcp.h return atomic_read(&sk->sk_rmem_alloc) > threshold; atomic_read 1329 include/net/xfrm.h return atomic_read(&x->tunnel_users); atomic_read 2878 include/rdma/ib_verbs.h if (atomic_read(usecnt) && ib_is_destroy_retryable(-EBUSY, why, uobj)) atomic_read 195 include/sound/hdaudio.h return atomic_read(&codec->in_pm); atomic_read 481 include/trace/events/bcache.h __entry->blocked = atomic_read(&ca->set->prio_blocked); atomic_read 1003 include/trace/events/btrfs.h __entry->refs = atomic_read(&buf->refs); atomic_read 190 include/trace/events/filelock.h __entry->wcount = atomic_read(&inode->i_writecount); atomic_read 191 include/trace/events/filelock.h __entry->rcount = atomic_read(&inode->i_readcount); atomic_read 192 include/trace/events/filelock.h __entry->icount = atomic_read(&inode->i_count); atomic_read 184 include/trace/events/fscache.h __entry->n_children = atomic_read(&cookie->n_children); atomic_read 185 include/trace/events/fscache.h __entry->n_active = atomic_read(&cookie->n_active); atomic_read 233 include/trace/events/fscache.h __entry->p_usage = atomic_read(&cookie->parent->usage); atomic_read 234 include/trace/events/fscache.h __entry->p_n_children = atomic_read(&cookie->parent->n_children); atomic_read 263 include/trace/events/fscache.h __entry->usage = atomic_read(&cookie->usage); atomic_read 264 include/trace/events/fscache.h __entry->n_children = atomic_read(&cookie->n_children); atomic_read 265 include/trace/events/fscache.h __entry->n_active = atomic_read(&cookie->n_active); atomic_read 291 include/trace/events/fscache.h __entry->usage = atomic_read(&cookie->usage); atomic_read 292 include/trace/events/fscache.h __entry->n_children = atomic_read(&cookie->n_children); atomic_read 293 include/trace/events/fscache.h __entry->n_active = atomic_read(&cookie->n_active); atomic_read 317 include/trace/events/fscache.h __entry->usage = atomic_read(&cookie->usage); atomic_read 318 include/trace/events/fscache.h __entry->n_children = atomic_read(&cookie->n_children); atomic_read 319 include/trace/events/fscache.h __entry->n_active = atomic_read(&cookie->n_active); atomic_read 84 include/trace/events/module.h __entry->refcnt = atomic_read(&mod->refcnt); atomic_read 47 include/trace/events/neigh.h __entry->entries = atomic_read(&tbl->gc_entries); atomic_read 1818 include/trace/events/rpcrdma.h __entry->avail = atomic_read(&rdma->sc_sq_avail); atomic_read 1854 include/trace/events/rpcrdma.h __entry->avail = atomic_read(&rdma->sc_sq_avail); atomic_read 38 include/trace/events/rpm.h __entry->usage_count = atomic_read( atomic_read 44 include/trace/events/rpm.h __entry->child_count = atomic_read( atomic_read 593 include/trace/events/rxrpc.h __entry->usage = atomic_read(&conn->usage); atomic_read 83 include/trace/events/sock.h __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc); atomic_read 115 include/trace/events/sock.h __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc); atomic_read 29 include/trace/events/vb2.h atomic_read(&q->owned_by_drv_count); atomic_read 479 ipc/msg.c msginfo->msgmap = atomic_read(&ns->msg_hdrs); atomic_read 480 ipc/msg.c msginfo->msgtql = atomic_read(&ns->msg_bytes); atomic_read 179 kernel/async.c if (!entry || atomic_read(&entry_count) > MAX_WORK) { atomic_read 373 kernel/audit.c atomic_read(&audit_lost), atomic_read 1202 kernel/audit.c s.lost = atomic_read(&audit_lost); atomic_read 674 kernel/cgroup/cgroup-v1.c atomic_read(&ss->root->nr_cgrps), atomic_read 1327 kernel/cgroup/cgroup.c BUG_ON(atomic_read(&root->nr_cgrps)); atomic_read 2073 kernel/cgroup/cgroup.c BUG_ON(atomic_read(&root->nr_cgrps) != 1); atomic_read 4558 kernel/cgroup/cgroup.c !atomic_read(&task->signal->live)) atomic_read 229 kernel/cgroup/debug.c atomic_read(&css->online_cnt), pbuf); atomic_read 75 kernel/cred.c return atomic_read(&cred->subscribers); atomic_read 101 kernel/cred.c atomic_read(&cred->usage) != 0 || atomic_read 106 kernel/cred.c atomic_read(&cred->usage), atomic_read 109 kernel/cred.c if (atomic_read(&cred->usage) != 0) atomic_read 111 kernel/cred.c cred, atomic_read(&cred->usage)); atomic_read 135 kernel/cred.c atomic_read(&cred->usage), atomic_read 138 kernel/cred.c BUG_ON(atomic_read(&cred->usage) != 0); atomic_read 162 kernel/cred.c atomic_read(&tsk->cred->usage), atomic_read 349 kernel/cred.c p->cred, atomic_read(&p->cred->usage), atomic_read 440 kernel/cred.c atomic_read(&new->usage), atomic_read 449 kernel/cred.c BUG_ON(atomic_read(&new->usage) < 1); atomic_read 523 kernel/cred.c atomic_read(&new->usage), atomic_read 529 kernel/cred.c BUG_ON(atomic_read(&new->usage) < 1); atomic_read 546 kernel/cred.c atomic_read(&new->usage), atomic_read 569 kernel/cred.c atomic_read(&old->usage), atomic_read 587 kernel/cred.c atomic_read(&old->usage), atomic_read 812 kernel/cred.c atomic_read(&cred->usage), atomic_read 886 kernel/cred.c atomic_read(&tsk->cred->usage), atomic_read 459 kernel/debug/debug_core.c if (atomic_read(&kgdb_setting_breakpoint)) atomic_read 476 kernel/debug/debug_core.c if (atomic_read(&kgdb_active) != raw_smp_processor_id()) atomic_read 615 kernel/debug/debug_core.c if (atomic_read(&kgdb_cpu_doing_single_step) != -1 && atomic_read 663 kernel/debug/debug_core.c (atomic_read(&masters_in_kgdb) + atomic_read(&slaves_in_kgdb)) != atomic_read 712 kernel/debug/debug_core.c while (kgdb_do_roundup && atomic_read(&slaves_in_kgdb)) atomic_read 717 kernel/debug/debug_core.c if (atomic_read(&kgdb_cpu_doing_single_step) != -1) { atomic_read 718 kernel/debug/debug_core.c int sstep_cpu = atomic_read(&kgdb_cpu_doing_single_step); atomic_read 855 kernel/debug/debug_core.c if (!kgdb_connected || atomic_read(&kgdb_active) != -1 || dbg_kdb_mode) atomic_read 1011 kernel/debug/debug_core.c if (atomic_read(&kgdb_break_tasklet_var) || atomic_read 1012 kernel/debug/debug_core.c atomic_read(&kgdb_active) != -1 || atomic_read 1013 kernel/debug/debug_core.c atomic_read(&kgdb_setting_breakpoint)) atomic_read 439 kernel/debug/gdbstub.c tid = -atomic_read(&kgdb_active) - 2; atomic_read 39 kernel/debug/kdb/kdb_debugger.c kdb_initial_cpu = atomic_read(&kgdb_active); atomic_read 69 kernel/debug/kdb/kdb_debugger.c if (atomic_read(&kgdb_setting_breakpoint)) atomic_read 2181 kernel/debug/kdb/kdb_main.c if (atomic_read(&kdb_nmi_disabled)) atomic_read 251 kernel/events/callchain.c if (atomic_read(&nr_callchain_events)) atomic_read 3326 kernel/events/core.c if (atomic_read(&nr_switch_events)) atomic_read 3337 kernel/events/core.c if (atomic_read(this_cpu_ptr(&perf_cgroup_events))) atomic_read 3576 kernel/events/core.c if (atomic_read(this_cpu_ptr(&perf_cgroup_events))) atomic_read 3587 kernel/events/core.c if (atomic_read(&nr_switch_events)) atomic_read 5634 kernel/events/core.c if (atomic_read(&rb->mmap_count)) atomic_read 7089 kernel/events/core.c if (!atomic_read(&nr_comm_events) && atomic_read 7090 kernel/events/core.c !atomic_read(&nr_mmap_events) && atomic_read 7091 kernel/events/core.c !atomic_read(&nr_task_events)) atomic_read 7200 kernel/events/core.c if (!atomic_read(&nr_comm_events)) atomic_read 7298 kernel/events/core.c if (!atomic_read(&nr_namespaces_events)) atomic_read 7667 kernel/events/core.c if (!atomic_read(&nr_mmap_events)) atomic_read 7948 kernel/events/core.c if (!atomic_read(&nr_ksymbol_events)) atomic_read 8067 kernel/events/core.c if (atomic_read(&nr_ksymbol_events)) atomic_read 8074 kernel/events/core.c if (!atomic_read(&nr_bpf_events)) atomic_read 8186 kernel/events/core.c int events = atomic_read(&event->event_limit); atomic_read 10393 kernel/events/core.c if (!atomic_read(&perf_sched_count)) { atomic_read 10794 kernel/events/core.c if (atomic_read(&event->mmap_count)) atomic_read 390 kernel/events/ring_buffer.c if (!atomic_read(&rb->aux_mmap_count)) atomic_read 1436 kernel/events/uprobes.c if (!atomic_read(&vma->vm_mm->mm_users)) /* called by mmput() ? */ atomic_read 1604 kernel/events/uprobes.c wait_event(area->wq, (atomic_read(&area->slot_count) < UINSNS_PER_PAGE)); atomic_read 363 kernel/exit.c if (atomic_read(&mm->mm_users) <= 1) { atomic_read 1073 kernel/fork.c VM_BUG_ON(atomic_read(&mm->mm_users)); atomic_read 1300 kernel/fork.c atomic_read(&mm->mm_users) > 1) { atomic_read 1872 kernel/fork.c if (atomic_read(&p->real_cred->user->processes) >= atomic_read 2792 kernel/fork.c (fd && atomic_read(&fd->count) > 1)) { atomic_read 373 kernel/futex.c return atomic_read(&hb->waiters); atomic_read 104 kernel/irq/manage.c return !atomic_read(&desc->threads_active); atomic_read 138 kernel/irq/manage.c !atomic_read(&desc->threads_active)); atomic_read 337 kernel/irq/spurious.c handled = atomic_read(&desc->threads_handled); atomic_read 110 kernel/jump_label.c int n = atomic_read(&key->enabled); atomic_read 135 kernel/jump_label.c for (v = atomic_read(&key->enabled); v > 0; v = v1) { atomic_read 142 kernel/jump_label.c if (atomic_read(&key->enabled) == 0) { atomic_read 169 kernel/jump_label.c if (atomic_read(&key->enabled) > 0) { atomic_read 170 kernel/jump_label.c WARN_ON_ONCE(atomic_read(&key->enabled) != 1); atomic_read 175 kernel/jump_label.c if (atomic_read(&key->enabled) == 0) { atomic_read 200 kernel/jump_label.c if (atomic_read(&key->enabled) != 1) { atomic_read 201 kernel/jump_label.c WARN_ON_ONCE(atomic_read(&key->enabled) != 0); atomic_read 154 kernel/kmod.c atomic_read(&kmod_concurrent_max), atomic_read 821 kernel/locking/locktorture.c if (atomic_read(&cxt.n_lock_torture_errors)) atomic_read 58 kernel/locking/osq_lock.c if (atomic_read(&lock->tail) == curr && atomic_read 68 kernel/locking/qrwlock.c if (!atomic_read(&lock->cnts) && atomic_read 221 kernel/locking/qspinlock.c u32 old, new, val = atomic_read(&lock->val); atomic_read 88 kernel/locking/qspinlock_paravirt.h int val = atomic_read(&lock->val); atomic_read 133 kernel/locking/qspinlock_paravirt.h int val = atomic_read(&lock->val); atomic_read 485 kernel/locking/qspinlock_paravirt.h return (u32)(atomic_read(&lock->val) | _Q_LOCKED_VAL); atomic_read 500 kernel/locking/qspinlock_paravirt.h (unsigned long)lock, atomic_read(&lock->val)); atomic_read 963 kernel/module.c return atomic_read(&mod->refcnt) - MODULE_REF_BASE; atomic_read 137 kernel/padata.c if (atomic_read(&pd->refcnt) >= MAX_OBJ_NUM) atomic_read 297 kernel/power/swap.c wait_event(hb->wait, atomic_read(&hb->count) == 0); atomic_read 597 kernel/power/swap.c wait_event(d->go, atomic_read(&d->ready) || atomic_read 640 kernel/power/swap.c wait_event(d->go, atomic_read(&d->ready) || atomic_read 805 kernel/power/swap.c atomic_read(&data[thr].stop)); atomic_read 844 kernel/power/swap.c wait_event(crc->done, atomic_read(&crc->stop)); atomic_read 1124 kernel/power/swap.c wait_event(d->go, atomic_read(&d->ready) || atomic_read 1330 kernel/power/swap.c wait_event(crc->done, atomic_read(&crc->stop)); atomic_read 1385 kernel/power/swap.c atomic_read(&data[thr].stop)); atomic_read 1430 kernel/power/swap.c wait_event(crc->done, atomic_read(&crc->stop)); atomic_read 75 kernel/printk/printk_safe.c len = atomic_read(&s->len); atomic_read 200 kernel/printk/printk_safe.c len = atomic_read(&s->len); atomic_read 266 kernel/rcu/rcu.h if (!atomic_read(&___rfd_beenhere) && \ atomic_read 311 kernel/rcu/rcuperf.c if (atomic_read(&n_rcu_perf_writer_finished) < nrealwriters) atomic_read 406 kernel/rcu/rcuperf.c if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { atomic_read 430 kernel/rcu/rcuperf.c atomic_read(&n_rcu_perf_writer_started) >= nrealwriters) atomic_read 459 kernel/rcu/rcuperf.c atomic_read(&n_rcu_perf_writer_finished) >= nrealwriters) atomic_read 594 kernel/rcu/rcuperf.c atomic_read(&n_rcu_perf_writer_finished) >= atomic_read 596 kernel/rcu/rcuperf.c } while (atomic_read(&n_rcu_perf_writer_finished) < nrealwriters); atomic_read 665 kernel/rcu/rcuperf.c while (atomic_read(&n_rcu_perf_reader_started) < nrealreaders) atomic_read 1426 kernel/rcu/rcutorture.c atomic_read(&n_rcu_torture_alloc), atomic_read 1427 kernel/rcu/rcutorture.c atomic_read(&n_rcu_torture_alloc_fail), atomic_read 1428 kernel/rcu/rcutorture.c atomic_read(&n_rcu_torture_free)); atomic_read 1430 kernel/rcu/rcutorture.c atomic_read(&n_rcu_torture_mberror), atomic_read 1445 kernel/rcu/rcutorture.c if (atomic_read(&n_rcu_torture_mberror) != 0 || atomic_read 1469 kernel/rcu/rcutorture.c pr_cont(" %d", atomic_read(&rcu_torture_wcount[i])); atomic_read 2058 kernel/rcu/rcutorture.c atomic_read(&barrier_cbs_count) == 0 || atomic_read 2064 kernel/rcu/rcutorture.c if (atomic_read(&barrier_cbs_invoked) != n_barrier_cbs) { atomic_read 2067 kernel/rcu/rcutorture.c atomic_read(&barrier_cbs_invoked), atomic_read 2249 kernel/rcu/rcutorture.c if (atomic_read(&n_rcu_torture_error) || n_rcu_torture_barrier_error) atomic_read 287 kernel/rcu/tree.c if (atomic_read(&rdp->dynticks) & RCU_DYNTICK_CTRL_CTR) atomic_read 301 kernel/rcu/tree.c return !(atomic_read(&rdp->dynticks) & RCU_DYNTICK_CTRL_CTR); atomic_read 348 kernel/rcu/tree.c old = atomic_read(&rdp->dynticks); atomic_read 579 kernel/rcu/tree.c trace_rcu_dyntick(TPS("Start"), rdp->dynticks_nesting, 0, atomic_read(&rdp->dynticks)); atomic_read 653 kernel/rcu/tree.c atomic_read(&rdp->dynticks)); atomic_read 660 kernel/rcu/tree.c trace_rcu_dyntick(TPS("Startirq"), rdp->dynticks_nmi_nesting, 0, atomic_read(&rdp->dynticks)); atomic_read 747 kernel/rcu/tree.c trace_rcu_dyntick(TPS("End"), rdp->dynticks_nesting, 1, atomic_read(&rdp->dynticks)); atomic_read 831 kernel/rcu/tree.c rdp->dynticks_nmi_nesting + incby, atomic_read(&rdp->dynticks)); atomic_read 2838 kernel/rcu/tree.c atomic_read(&rcu_state.barrier_cpu_count), done); atomic_read 1539 kernel/rcu/tree_plugin.h while (WARN_ON_ONCE(atomic_read(&rdp->nocb_lock_contended))) atomic_read 2427 kernel/rcu/tree_plugin.h "cC"[!!atomic_read(&rdp->nocb_lock_contended)], atomic_read 619 kernel/rcu/tree_stall.h atomic_read(&warned)) atomic_read 628 kernel/rcu/tree_stall.h atomic_read(&warned)) { atomic_read 155 kernel/rcu/update.c return rcu_expedited || atomic_read(&rcu_expedited_nesting); atomic_read 638 kernel/sched/core.c if (!(atomic_read(nohz_flags(cpu)) & NOHZ_KICK_MASK)) atomic_read 3441 kernel/sched/core.c return atomic_read(&cpu_rq(cpu)->nr_iowait); atomic_read 71 kernel/sched/cpupri.c if (!atomic_read(&(vec)->count)) atomic_read 223 kernel/sched/cputime.c if (atomic_read(&rq->nr_iowait) > 0) atomic_read 380 kernel/sched/deadline.c return atomic_read(&rq->rd->dlo_count); atomic_read 2632 kernel/sched/fair.c mm_users = atomic_read(&mm->mm_users); atomic_read 9421 kernel/sched/fair.c if (likely(!atomic_read(&nohz.nr_cpus))) atomic_read 9498 kernel/sched/fair.c nr_busy = atomic_read(&sds->nr_busy_cpus); atomic_read 9734 kernel/sched/fair.c if (!(atomic_read(nohz_flags(this_cpu)) & NOHZ_KICK_MASK)) atomic_read 40 kernel/sched/membarrier.c atomic_read(&mm->membarrier_state)); atomic_read 141 kernel/sched/membarrier.c if (!(atomic_read(&mm->membarrier_state) & atomic_read 145 kernel/sched/membarrier.c if (!(atomic_read(&mm->membarrier_state) & atomic_read 150 kernel/sched/membarrier.c if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) atomic_read 202 kernel/sched/membarrier.c int membarrier_state = atomic_read(&mm->membarrier_state); atomic_read 206 kernel/sched/membarrier.c if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) { atomic_read 265 kernel/sched/membarrier.c if (atomic_read(&mm->membarrier_state) & atomic_read 298 kernel/sched/membarrier.c if ((atomic_read(&mm->membarrier_state) & ready_state) == ready_state) atomic_read 271 kernel/sched/rt.c return atomic_read(&rq->rd->rto_count); atomic_read 2492 kernel/sched/sched.h membarrier_state = atomic_read(&next_mm->membarrier_state); atomic_read 1225 kernel/sched/topology.c if (!atomic_read(&d->rd->refcount)) atomic_read 1268 kernel/sched/topology.c if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) atomic_read 1271 kernel/sched/topology.c if (atomic_read(&(*per_cpu_ptr(sdd->sg, cpu))->ref)) atomic_read 1274 kernel/sched/topology.c if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref)) atomic_read 337 kernel/smpboot.c return atomic_read(&per_cpu(cpu_hotplug_state, cpu)); atomic_read 359 kernel/smpboot.c switch (atomic_read(&per_cpu(cpu_hotplug_state, cpu))) { atomic_read 428 kernel/smpboot.c if (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) == CPU_DEAD) atomic_read 433 kernel/smpboot.c while (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) != CPU_DEAD) { atomic_read 441 kernel/smpboot.c oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu)); atomic_read 472 kernel/smpboot.c oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu)); atomic_read 519 kernel/softirq.c if (!atomic_read(&t->count)) { atomic_read 467 kernel/sys.c if (atomic_read(&new_user->processes) >= rlimit(RLIMIT_NPROC) && atomic_read 198 kernel/time/clocksource.c reset_pending = atomic_read(&watchdog_reset_pending); atomic_read 216 kernel/time/clocksource.c atomic_read(&watchdog_reset_pending)) { atomic_read 234 kernel/time/clocksource.c if (atomic_read(&watchdog_reset_pending)) atomic_read 1204 kernel/time/hrtimer.c if (atomic_read(&cpu_base->timer_waiters)) { atomic_read 183 kernel/time/tick-sched.c int val = atomic_read(dep); atomic_read 389 kernel/time/tick-sched.c if (atomic_read(¤t->tick_dep_mask) || atomic_read 390 kernel/time/tick-sched.c atomic_read(¤t->signal->tick_dep_mask)) atomic_read 1259 kernel/time/timer.c if (atomic_read(&base->timer_waiters)) { atomic_read 376 kernel/trace/blktrace.c snprintf(buf, sizeof(buf), "%u\n", atomic_read(&bt->dropped)); atomic_read 170 kernel/trace/fgraph.c trace->overrun = atomic_read(¤t->trace_overrun); atomic_read 1746 kernel/trace/ring_buffer.c if (atomic_read(&buffer->resize_disabled)) atomic_read 1854 kernel/trace/ring_buffer.c if (atomic_read(&buffer->record_disabled)) { atomic_read 2973 kernel/trace/ring_buffer.c if (unlikely(atomic_read(&buffer->record_disabled))) atomic_read 2983 kernel/trace/ring_buffer.c if (unlikely(atomic_read(&cpu_buffer->record_disabled))) atomic_read 3128 kernel/trace/ring_buffer.c if (atomic_read(&buffer->record_disabled)) atomic_read 3138 kernel/trace/ring_buffer.c if (atomic_read(&cpu_buffer->record_disabled)) atomic_read 3232 kernel/trace/ring_buffer.c rd = atomic_read(&buffer->record_disabled); atomic_read 3255 kernel/trace/ring_buffer.c rd = atomic_read(&buffer->record_disabled); atomic_read 3269 kernel/trace/ring_buffer.c return !atomic_read(&buffer->record_disabled); atomic_read 3285 kernel/trace/ring_buffer.c return !(atomic_read(&buffer->record_disabled) & RB_BUFFER_OFF); atomic_read 4533 kernel/trace/ring_buffer.c if (atomic_read(&buffer_a->record_disabled)) atomic_read 4536 kernel/trace/ring_buffer.c if (atomic_read(&buffer_b->record_disabled)) atomic_read 4539 kernel/trace/ring_buffer.c if (atomic_read(&cpu_buffer_a->record_disabled)) atomic_read 4542 kernel/trace/ring_buffer.c if (atomic_read(&cpu_buffer_b->record_disabled)) atomic_read 2249 kernel/trace/trace.c if (atomic_read(&trace_record_taskinfo_disabled) || !tracing_is_on()) atomic_read 59 kernel/trace/trace_branch.c if (atomic_read(&data->disabled)) atomic_read 1151 kernel/trace/trace_events_hist.c if (unlikely(atomic_read(&tp->key.enabled) > 0)) { atomic_read 147 kernel/trace/trace_functions.c if (!atomic_read(&data->disabled)) { atomic_read 387 kernel/trace/trace_irqsoff.c if (unlikely(!data) || atomic_read(&data->disabled)) atomic_read 426 kernel/trace/trace_irqsoff.c !data->critical_start || atomic_read(&data->disabled)) atomic_read 178 kernel/ucount.c c = atomic_read(v); atomic_read 335 kernel/umh.c atomic_read(&running_helpers) == 0, atomic_read 770 kernel/workqueue.c return !atomic_read(&pool->nr_running); atomic_read 796 kernel/workqueue.c atomic_read(&pool->nr_running) <= 1; atomic_read 1796 kernel/workqueue.c atomic_read(&pool->nr_running)); atomic_read 2731 kernel/workqueue.c WARN_ON_ONCE(atomic_read(&wq->nr_pwqs_to_flush)); atomic_read 114 lib/dump_stack.c do { cpu_relax(); } while (atomic_read(&dump_lock) != -1); atomic_read 49 lib/fault-inject.c atomic_read(&attr->space), atomic_read 50 lib/fault-inject.c atomic_read(&attr->times)); atomic_read 123 lib/fault-inject.c if (atomic_read(&attr->times) == 0) atomic_read 126 lib/fault-inject.c if (atomic_read(&attr->space) > size) { atomic_read 146 lib/fault-inject.c if (atomic_read(&attr->times) != -1) atomic_read 22 lib/is_single_threaded.c if (atomic_read(&task->signal->live) != 1) atomic_read 25 lib/is_single_threaded.c if (atomic_read(&mm->mm_users) == 1) atomic_read 66 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); atomic_read 123 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); atomic_read 182 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); atomic_read 277 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); atomic_read 381 lib/rhashtable.c unsigned int nelems = atomic_read(&ht->nelems); atomic_read 509 lib/sbitmap.c if (!atomic_read(&sbq->ws_active)) atomic_read 512 lib/sbitmap.c wake_index = atomic_read(&sbq->wake_index); atomic_read 517 lib/sbitmap.c if (wake_index != atomic_read(&sbq->wake_index)) atomic_read 615 lib/sbitmap.c wake_index = atomic_read(&sbq->wake_index); atomic_read 645 lib/sbitmap.c seq_printf(m, "wake_index=%d\n", atomic_read(&sbq->wake_index)); atomic_read 646 lib/sbitmap.c seq_printf(m, "ws_active=%d\n", atomic_read(&sbq->ws_active)); atomic_read 653 lib/sbitmap.c atomic_read(&ws->wait_cnt), atomic_read 202 lib/test_rhashtable.c total, atomic_read(&ht->nelems), entries, chain_len); atomic_read 204 lib/test_rhashtable.c if (total != atomic_read(&ht->nelems) || total != entries) atomic_read 638 lib/test_rhashtable.c if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == -1)) { atomic_read 793 lib/test_rhashtable.c if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == 0)) atomic_read 1136 mm/backing-dev.c if (atomic_read(&nr_wb_congested[sync]) == 0) { atomic_read 171 mm/debug.c mm->pgd, atomic_read(&mm->mm_users), atomic_read 172 mm/debug.c atomic_read(&mm->mm_count), atomic_read 195 mm/debug.c atomic_read(&mm->tlb_flush_pending), atomic_read 378 mm/frontswap.c totalpages += atomic_read(&si->frontswap_pages); atomic_read 393 mm/frontswap.c si_frontswap_pages = atomic_read(&si->frontswap_pages); atomic_read 140 mm/huge_memory.c return atomic_read(&huge_zero_refcount) == 1 ? HPAGE_PMD_NR : 0; atomic_read 2441 mm/huge_memory.c VM_BUG_ON_PAGE(atomic_read(&page_tail->_mapcount) != -1, page_tail); atomic_read 2586 mm/huge_memory.c return atomic_read(&page->_mapcount) + 1; atomic_read 2593 mm/huge_memory.c ret += atomic_read(&page[i]._mapcount) + 1; atomic_read 2634 mm/huge_memory.c mapcount = atomic_read(&page->_mapcount) + 1; atomic_read 2644 mm/huge_memory.c mapcount = atomic_read(&page[i]._mapcount) + 1; atomic_read 404 mm/khugepaged.c return atomic_read(&mm->mm_users) == 0; atomic_read 664 mm/kmemleak.c WARN_ON(atomic_read(&object->use_count) < 1); atomic_read 1416 mm/kmemleak.c if (atomic_read(&object->use_count) > 1) { atomic_read 1418 mm/kmemleak.c atomic_read(&object->use_count)); atomic_read 452 mm/ksm.c return atomic_read(&mm->mm_users) == 0; atomic_read 1653 mm/memcontrol.c if (!atomic_read(&memcg->numainfo_events)) atomic_read 2111 mm/memcontrol.c if (atomic_read(&memcg->moving_account) <= 0) atomic_read 4529 mm/memcontrol.c atomic_read(&frn->done.cnt) == 1) { atomic_read 4576 mm/memcontrol.c atomic_read(&frn->done.cnt) == 1) { atomic_read 765 mm/migrate.c if (atomic_read(&bh->b_count)) { atomic_read 255 mm/mmu_notifier.c BUG_ON(atomic_read(&mm->mm_users) <= 0); atomic_read 305 mm/mmu_notifier.c BUG_ON(atomic_read(&mm->mm_users) <= 0); atomic_read 430 mm/mmu_notifier.c BUG_ON(atomic_read(&mm->mm_count) <= 0); atomic_read 463 mm/mmu_notifier.c BUG_ON(atomic_read(&mm->mm_count) <= 0); atomic_read 65 mm/mprotect.c atomic_read(&vma->vm_mm->mm_users) == 1) atomic_read 771 mm/oom_kill.c !atomic_read(&oom_victims), timeout); atomic_read 833 mm/oom_kill.c if (atomic_read(&mm->mm_users) <= 1) atomic_read 1014 mm/page_alloc.c if (unlikely(atomic_read(&page->_mapcount) != -1)) atomic_read 1036 mm/page_alloc.c if (unlikely(atomic_read(&page->_mapcount) != -1)) atomic_read 2048 mm/page_alloc.c if (unlikely(atomic_read(&page->_mapcount) != -1)) atomic_read 100 mm/rmap.c VM_BUG_ON(atomic_read(&anon_vma->refcount)); atomic_read 1104 mm/shmem.c !atomic_read(&info->stop_eviction)); atomic_read 1107 mm/shmem.c if (!atomic_read(&info->stop_eviction)) atomic_read 4094 mm/slab.c unsigned long allochit = atomic_read(&cachep->allochit); atomic_read 4095 mm/slab.c unsigned long allocmiss = atomic_read(&cachep->allocmiss); atomic_read 4096 mm/slab.c unsigned long freehit = atomic_read(&cachep->freehit); atomic_read 4097 mm/slab.c unsigned long freemiss = atomic_read(&cachep->freemiss); atomic_read 301 mm/swap_state.c return READ_ONCE(enable_vma_readahead) && !atomic_read(&nr_rotate_swap); atomic_read 513 mm/swap_state.c atomic_read(&last_readahead_pages)); atomic_read 1619 mm/swapfile.c mapcount = atomic_read(&page[i]._mapcount) + 1; atomic_read 2700 mm/swapfile.c if (seq->poll_event != atomic_read(&proc_poll_event)) { atomic_read 2701 mm/swapfile.c seq->poll_event = atomic_read(&proc_poll_event); atomic_read 2795 mm/swapfile.c seq->poll_event = atomic_read(&proc_poll_event); atomic_read 641 mm/util.c return atomic_read(&page->_mapcount) >= 0; atomic_read 643 mm/util.c if (atomic_read(compound_mapcount_ptr(page)) >= 0) atomic_read 648 mm/util.c if (atomic_read(&page[i]._mapcount) >= 0) atomic_read 706 mm/util.c ret = atomic_read(&page->_mapcount) + 1; atomic_read 714 mm/util.c ret += atomic_read(compound_mapcount_ptr(page)) + 1; atomic_read 64 mm/zpool.c refcount = atomic_read(&driver->refcount); atomic_read 430 net/9p/trans_rdma.c if (unlikely(atomic_read(&rdma->excess_rc) > 0)) { atomic_read 323 net/9p/trans_virtio.c if (atomic_read(&vp_pinned) >= chan->p9_max_pages) { atomic_read 325 net/9p/trans_virtio.c (atomic_read(&vp_pinned) < chan->p9_max_pages)); atomic_read 18 net/atm/atm_misc.c if (atomic_read(&sk_atm(vcc)->sk_rmem_alloc) <= sk_atm(vcc)->sk_rcvbuf) atomic_read 33 net/atm/atm_misc.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) { atomic_read 90 net/atm/atm_misc.c #define __HANDLE_ITEM(i) to->i = atomic_read(&from->i) atomic_read 264 net/atm/br2684.c if (unlikely(atomic_read(&brvcc->qspace) > 0)) atomic_read 278 net/atm/br2684.c if (atomic_read(&brvcc->qspace) > 0) atomic_read 80 net/atm/common.c if (atomic_read(&sk->sk_rmem_alloc)) atomic_read 82 net/atm/common.c __func__, atomic_read(&sk->sk_rmem_alloc)); atomic_read 559 net/atm/common.c pr_debug("%d -= %d\n", atomic_read(&sk->sk_rmem_alloc), atomic_read 48 net/atm/proc.c atomic_read(&stats->tx), atomic_read(&stats->tx_err), atomic_read 49 net/atm/proc.c atomic_read(&stats->rx), atomic_read(&stats->rx_err), atomic_read 50 net/atm/proc.c atomic_read(&stats->rx_drop)); atomic_read 164 net/atm/resources.c #define __HANDLE_ITEM(i) to->i = atomic_read(&from->i) atomic_read 250 net/ax25/af_ax25.c atomic_read(&s->sk->sk_rmem_alloc) <= s->sk->sk_rcvbuf) { atomic_read 124 net/ax25/ax25_ds_timer.c if (atomic_read(&sk->sk_rmem_alloc) < atomic_read 263 net/ax25/ax25_in.c if (atomic_read(&sk->sk_rmem_alloc) >= atomic_read 62 net/ax25/ax25_std_timer.c if (atomic_read(&sk->sk_rmem_alloc) < atomic_read 282 net/batman-adv/bat_iv_ogm.c msecs = atomic_read(&bat_priv->orig_interval) - BATADV_JITTER; atomic_read 297 net/batman-adv/bat_iv_ogm.c int hop_penalty = atomic_read(&bat_priv->hop_penalty); atomic_read 550 net/batman-adv/bat_iv_ogm.c if (atomic_read(&bat_priv->aggregated_ogms) && atomic_read 641 net/batman-adv/bat_iv_ogm.c if (atomic_read(&bat_priv->aggregated_ogms) && !own_packet) { atomic_read 667 net/batman-adv/bat_iv_ogm.c if (!own_packet && atomic_read(&bat_priv->aggregated_ogms)) atomic_read 821 net/batman-adv/bat_iv_ogm.c seqno = (u32)atomic_read(&hard_iface->bat_iv.ogm_seqno); atomic_read 1592 net/batman-adv/bat_iv_ogm.c if_incoming_seqno = atomic_read(&if_incoming->bat_iv.ogm_seqno); atomic_read 1705 net/batman-adv/bat_iv_ogm.c if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING) { atomic_read 2438 net/batman-adv/bat_iv_ogm.c switch (atomic_read(&bat_priv->gw.sel_class)) { atomic_read 2501 net/batman-adv/bat_iv_ogm.c if (atomic_read(&bat_priv->gw.sel_class) <= 2) atomic_read 2536 net/batman-adv/bat_iv_ogm.c if ((atomic_read(&bat_priv->gw.sel_class) > 3) && atomic_read 2537 net/batman-adv/bat_iv_ogm.c (orig_tq_avg - gw_tq_avg < atomic_read(&bat_priv->gw.sel_class))) atomic_read 679 net/batman-adv/bat_v.c old_class = atomic_read(&bat_priv->gw.sel_class); atomic_read 690 net/batman-adv/bat_v.c u32 class = atomic_read(&bat_priv->gw.sel_class); atomic_read 792 net/batman-adv/bat_v.c threshold = atomic_read(&bat_priv->gw.sel_class); atomic_read 51 net/batman-adv/bat_v_elp.c msecs = atomic_read(&hard_iface->bat_v.elp_interval) - BATADV_JITTER; atomic_read 77 net/batman-adv/bat_v_elp.c throughput = atomic_read(&hard_iface->bat_v.throughput_override); atomic_read 270 net/batman-adv/bat_v_elp.c if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING) atomic_read 287 net/batman-adv/bat_v_elp.c elp_packet->seqno = htonl(atomic_read(&hard_iface->bat_v.elp_seqno)); atomic_read 288 net/batman-adv/bat_v_elp.c elp_interval = atomic_read(&hard_iface->bat_v.elp_interval); atomic_read 294 net/batman-adv/bat_v_elp.c atomic_read(&hard_iface->bat_v.elp_seqno)); atomic_read 109 net/batman-adv/bat_v_ogm.c msecs = atomic_read(&bat_priv->orig_interval) - BATADV_JITTER; atomic_read 245 net/batman-adv/bat_v_ogm.c if (!atomic_read(&bat_priv->aggregated_ogms)) { atomic_read 275 net/batman-adv/bat_v_ogm.c if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING) atomic_read 299 net/batman-adv/bat_v_ogm.c ogm_packet->seqno = htonl(atomic_read(&bat_priv->bat_v.ogm_seqno)); atomic_read 479 net/batman-adv/bat_v_ogm.c int hop_penalty = atomic_read(&bat_priv->hop_penalty); atomic_read 646 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&backbone_gw->request_sent)) { atomic_read 869 net/batman-adv/bridge_loop_avoidance.c if (atomic_read(&backbone_gw->request_sent)) { atomic_read 1245 net/batman-adv/bridge_loop_avoidance.c if (atomic_read(&backbone_gw->request_sent)) atomic_read 1337 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&bat_priv->bridge_loop_avoidance)) atomic_read 1441 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&bat_priv->bridge_loop_avoidance)) atomic_read 1489 net/batman-adv/bridge_loop_avoidance.c if (atomic_read(&backbone_gw->request_sent) == 0) atomic_read 1668 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&bat_priv->bridge_loop_avoidance)) atomic_read 1706 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&orig_node->bat_priv->bridge_loop_avoidance)) atomic_read 1839 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&bat_priv->bridge_loop_avoidance)) atomic_read 1845 net/batman-adv/bridge_loop_avoidance.c if (unlikely(atomic_read(&bat_priv->bla.num_requests))) atomic_read 1950 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&bat_priv->bridge_loop_avoidance)) atomic_read 1958 net/batman-adv/bridge_loop_avoidance.c if (unlikely(atomic_read(&bat_priv->bla.num_requests))) atomic_read 2177 net/batman-adv/bridge_loop_avoidance.c cb->seq = atomic_read(&hash->generation) << 1 | 1; atomic_read 2415 net/batman-adv/bridge_loop_avoidance.c cb->seq = atomic_read(&hash->generation) << 1 | 1; atomic_read 2518 net/batman-adv/bridge_loop_avoidance.c if (!atomic_read(&bat_priv->bridge_loop_avoidance)) atomic_read 744 net/batman-adv/distributed-arp-table.c dat_mode = atomic_read(&bat_priv->distributed_arp_table); atomic_read 961 net/batman-adv/distributed-arp-table.c cb->seq = atomic_read(&hash->generation) << 1 | 1; atomic_read 1204 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 1300 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 1363 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 1411 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 1782 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 1812 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 1857 net/batman-adv/distributed-arp-table.c if (!atomic_read(&bat_priv->distributed_arp_table)) atomic_read 186 net/batman-adv/gateway_client.c if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT) atomic_read 218 net/batman-adv/gateway_client.c if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT) atomic_read 768 net/batman-adv/gateway_client.c switch (atomic_read(&bat_priv->gw.mode)) { atomic_read 140 net/batman-adv/gateway_common.c gw_mode = atomic_read(&bat_priv->gw.mode); atomic_read 148 net/batman-adv/gateway_common.c down = atomic_read(&bat_priv->gw.bandwidth_down); atomic_read 149 net/batman-adv/gateway_common.c up = atomic_read(&bat_priv->gw.bandwidth_up); atomic_read 177 net/batman-adv/gateway_common.c down_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_down); atomic_read 178 net/batman-adv/gateway_common.c up_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_up); atomic_read 246 net/batman-adv/gateway_common.c atomic_read(&bat_priv->gw.mode) == BATADV_GW_MODE_CLIENT) atomic_read 587 net/batman-adv/hard-interface.c if (atomic_read(&bat_priv->fragmentation) == 0) atomic_read 777 net/batman-adv/hard-interface.c if (atomic_read(&bat_priv->fragmentation) && atomic_read 784 net/batman-adv/hard-interface.c if (!atomic_read(&bat_priv->fragmentation) && atomic_read 230 net/batman-adv/icmp_socket.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 82 net/batman-adv/log.h if (atomic_read(&__batpriv->log_level) & (type) && \ atomic_read 485 net/batman-adv/main.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 682 net/batman-adv/main.c ap_isolation_enabled = atomic_read(&vlan->ap_isolation); atomic_read 1104 net/batman-adv/multicast.c if (!atomic_read(&bat_priv->multicast_mode)) atomic_read 1139 net/batman-adv/multicast.c return atomic_read(&bat_priv->mcast.num_want_all_ipv4); atomic_read 1141 net/batman-adv/multicast.c return atomic_read(&bat_priv->mcast.num_want_all_ipv6); atomic_read 1163 net/batman-adv/multicast.c return atomic_read(&bat_priv->mcast.num_want_all_rtr4); atomic_read 1165 net/batman-adv/multicast.c return atomic_read(&bat_priv->mcast.num_want_all_rtr6); atomic_read 1404 net/batman-adv/multicast.c atomic_read(&bat_priv->mcast.num_want_all_unsnoopables); atomic_read 1428 net/batman-adv/multicast.c mcast_fanout = atomic_read(&bat_priv->multicast_fanout); atomic_read 2259 net/batman-adv/multicast.c cb->seq = atomic_read(&hash->generation) << 1 | 1; atomic_read 187 net/batman-adv/netlink.c ap_isolation = atomic_read(&vlan->ap_isolation); atomic_read 249 net/batman-adv/netlink.c (u8)atomic_read(&bat_priv->tt.vn))) atomic_read 275 net/batman-adv/netlink.c !!atomic_read(&bat_priv->aggregated_ogms))) atomic_read 290 net/batman-adv/netlink.c !!atomic_read(&bat_priv->bonding))) atomic_read 295 net/batman-adv/netlink.c !!atomic_read(&bat_priv->bridge_loop_avoidance))) atomic_read 301 net/batman-adv/netlink.c !!atomic_read(&bat_priv->distributed_arp_table))) atomic_read 306 net/batman-adv/netlink.c !!atomic_read(&bat_priv->fragmentation))) atomic_read 310 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.bandwidth_down))) atomic_read 314 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.bandwidth_up))) atomic_read 318 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.mode))) atomic_read 327 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.sel_class))) atomic_read 332 net/batman-adv/netlink.c atomic_read(&bat_priv->hop_penalty))) atomic_read 337 net/batman-adv/netlink.c atomic_read(&bat_priv->log_level))) atomic_read 343 net/batman-adv/netlink.c !atomic_read(&bat_priv->multicast_mode))) atomic_read 347 net/batman-adv/netlink.c atomic_read(&bat_priv->multicast_fanout))) atomic_read 353 net/batman-adv/netlink.c !!atomic_read(&bat_priv->network_coding))) atomic_read 358 net/batman-adv/netlink.c atomic_read(&bat_priv->orig_interval))) atomic_read 831 net/batman-adv/netlink.c atomic_read(&hard_iface->bat_v.elp_interval))) atomic_read 835 net/batman-adv/netlink.c atomic_read(&hard_iface->bat_v.throughput_override))) atomic_read 1041 net/batman-adv/netlink.c !!atomic_read(&vlan->ap_isolation))) atomic_read 92 net/batman-adv/network-coding.c nc_mode = atomic_read(&bat_priv->network_coding); atomic_read 277 net/batman-adv/network-coding.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 293 net/batman-adv/network-coding.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 314 net/batman-adv/network-coding.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 604 net/batman-adv/network-coding.c if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_ACTIVE && atomic_read 642 net/batman-adv/network-coding.c if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_ACTIVE && atomic_read 907 net/batman-adv/network-coding.c if (!atomic_read(&bat_priv->network_coding)) atomic_read 1512 net/batman-adv/network-coding.c if (!atomic_read(&bat_priv->network_coding)) atomic_read 1565 net/batman-adv/network-coding.c if (!atomic_read(&bat_priv->network_coding)) atomic_read 1809 net/batman-adv/network-coding.c if (!atomic_read(&bat_priv->network_coding)) atomic_read 565 net/batman-adv/routing.c if (!(recv_if == BATADV_IF_DEFAULT && atomic_read(&bat_priv->bonding))) atomic_read 776 net/batman-adv/routing.c orig_ttvn = (u8)atomic_read(&bat_priv->tt.vn); atomic_read 787 net/batman-adv/routing.c orig_ttvn = (u8)atomic_read(&orig_node->last_ttvn); atomic_read 854 net/batman-adv/routing.c curr_ttvn = (u8)atomic_read(&bat_priv->tt.vn); atomic_read 865 net/batman-adv/routing.c curr_ttvn = (u8)atomic_read(&orig_node->last_ttvn); atomic_read 195 net/batman-adv/send.c if (atomic_read(&bat_priv->fragmentation) && atomic_read 239 net/batman-adv/send.c u8 ttvn = (u8)atomic_read(&orig_node->last_ttvn); atomic_read 871 net/batman-adv/send.c if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING) { atomic_read 142 net/batman-adv/soft-interface.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 206 net/batman-adv/soft-interface.c if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE) atomic_read 271 net/batman-adv/soft-interface.c gw_mode = atomic_read(&bat_priv->gw.mode); atomic_read 162 net/batman-adv/sysfs.c atomic_read(&bat_priv->_name) == 0 ? \ atomic_read 202 net/batman-adv/sysfs.c return sprintf(buff, "%i\n", atomic_read(&bat_priv->_var)); \ atomic_read 244 net/batman-adv/sysfs.c atomic_read(&vlan->_name) == 0 ? \ atomic_read 302 net/batman-adv/sysfs.c length = sprintf(buff, "%i\n", atomic_read(&hard_iface->_var)); \ atomic_read 346 net/batman-adv/sysfs.c if (atomic_read(attr) == enabled) atomic_read 350 net/batman-adv/sysfs.c atomic_read(attr) == 1 ? "enabled" : "disabled", atomic_read 406 net/batman-adv/sysfs.c if (atomic_read(attr) == uint_val) atomic_read 413 net/batman-adv/sysfs.c attr_name, ifname, atomic_read(attr), uint_val); atomic_read 468 net/batman-adv/sysfs.c switch (atomic_read(&bat_priv->gw.mode)) { atomic_read 526 net/batman-adv/sysfs.c if (atomic_read(&bat_priv->gw.mode) == gw_mode_tmp) atomic_read 529 net/batman-adv/sysfs.c switch (atomic_read(&bat_priv->gw.mode)) { atomic_read 583 net/batman-adv/sysfs.c return sprintf(buff, "%i\n", atomic_read(&bat_priv->gw.sel_class)); atomic_read 627 net/batman-adv/sysfs.c down = atomic_read(&bat_priv->gw.bandwidth_down); atomic_read 628 net/batman-adv/sysfs.c up = atomic_read(&bat_priv->gw.bandwidth_up); atomic_read 1155 net/batman-adv/sysfs.c old_tp_override = atomic_read(&hard_iface->bat_v.throughput_override); atomic_read 1191 net/batman-adv/sysfs.c tp_override = atomic_read(&hard_iface->bat_v.throughput_override); atomic_read 467 net/batman-adv/tp_meter.c if (unlikely(atomic_read(&tp_vars->sending) == 0)) atomic_read 487 net/batman-adv/tp_meter.c if (atomic_read(&tp_vars->sending) == 0) atomic_read 511 net/batman-adv/tp_meter.c atomic_read(&tp_vars->last_acked)); atomic_read 518 net/batman-adv/tp_meter.c tp_vars->last_sent = atomic_read(&tp_vars->last_acked); atomic_read 646 net/batman-adv/tp_meter.c if (unlikely(atomic_read(&tp_vars->sending) == 0)) atomic_read 651 net/batman-adv/tp_meter.c (u32)atomic_read(&tp_vars->last_acked))) atomic_read 673 net/batman-adv/tp_meter.c if (atomic_read(&tp_vars->last_acked) == recv_ack) { atomic_read 675 net/batman-adv/tp_meter.c if (atomic_read(&tp_vars->dup_acks) != 3) atomic_read 708 net/batman-adv/tp_meter.c atomic64_add(recv_ack - atomic_read(&tp_vars->last_acked), atomic_read 741 net/batman-adv/tp_meter.c if (recv_ack - atomic_read(&tp_vars->last_acked) >= mss) atomic_read 770 net/batman-adv/tp_meter.c win_limit = atomic_read(&tp_vars->last_acked) + tp_vars->cwnd; atomic_read 848 net/batman-adv/tp_meter.c while (atomic_read(&tp_vars->sending) != 0) { atomic_read 316 net/batman-adv/translation-table.c count = atomic_read(&tt_global_entry->orig_list_count); atomic_read 573 net/batman-adv/translation-table.c tt_local_entries += atomic_read(&vlan->tt.num_entries); atomic_read 703 net/batman-adv/translation-table.c packet_size_max = atomic_read(&bat_priv->packet_size_max); atomic_read 729 net/batman-adv/translation-table.c (u8)atomic_read(&bat_priv->tt.vn)); atomic_read 867 net/batman-adv/translation-table.c num_entries += atomic_read(&vlan->tt.num_entries); atomic_read 887 net/batman-adv/translation-table.c (*tt_data)->ttvn = atomic_read(&orig_node->last_ttvn); atomic_read 941 net/batman-adv/translation-table.c vlan_entries = atomic_read(&vlan->tt.num_entries); atomic_read 966 net/batman-adv/translation-table.c (*tt_data)->ttvn = atomic_read(&bat_priv->tt.vn); atomic_read 971 net/batman-adv/translation-table.c vlan_entries = atomic_read(&vlan->tt.num_entries); atomic_read 1004 net/batman-adv/translation-table.c tt_diff_entries_num = atomic_read(&bat_priv->tt.local_changes); atomic_read 1096 net/batman-adv/translation-table.c net_dev->name, (u8)atomic_read(&bat_priv->tt.vn)); atomic_read 1225 net/batman-adv/translation-table.c cb->seq = atomic_read(&hash->generation) << 1 | 1; atomic_read 1919 net/batman-adv/translation-table.c last_ttvn = atomic_read(&best_entry->orig_node->last_ttvn); atomic_read 1951 net/batman-adv/translation-table.c last_ttvn = atomic_read(&orig_entry->orig_node->last_ttvn); atomic_read 2053 net/batman-adv/translation-table.c last_ttvn = atomic_read(&orig->orig_node->last_ttvn); atomic_read 3257 net/batman-adv/translation-table.c orig_ttvn = (u8)atomic_read(&req_dst_orig_node->last_ttvn); atomic_read 3313 net/batman-adv/translation-table.c if (tt_len > atomic_read(&bat_priv->packet_size_max)) { atomic_read 3381 net/batman-adv/translation-table.c my_ttvn = (u8)atomic_read(&bat_priv->tt.vn); atomic_read 3420 net/batman-adv/translation-table.c req_ttvn = (u8)atomic_read(&bat_priv->tt.vn); atomic_read 3929 net/batman-adv/translation-table.c if (atomic_read(&bat_priv->tt.local_changes) < 1) { atomic_read 3944 net/batman-adv/translation-table.c (u8)atomic_read(&bat_priv->tt.vn)); atomic_read 3984 net/batman-adv/translation-table.c if (!atomic_read(&vlan->ap_isolation)) atomic_read 4023 net/batman-adv/translation-table.c u8 orig_ttvn = (u8)atomic_read(&orig_node->last_ttvn); atomic_read 4165 net/batman-adv/translation-table.c atomic_read(&orig_node->last_ttvn))) atomic_read 4186 net/batman-adv/translation-table.c int packet_size_max = atomic_read(&bat_priv->packet_size_max); atomic_read 4380 net/batman-adv/translation-table.c atomic_read(&orig_node->last_ttvn) + 1); atomic_read 110 net/bluetooth/6lowpan.c BT_DBG("peers %d addr %pMR type %d", atomic_read(&dev->peer_count), atomic_read 164 net/bluetooth/6lowpan.c int count = atomic_read(&dev->peer_count); atomic_read 821 net/bluetooth/6lowpan.c if (!err && last && dev && !atomic_read(&dev->peer_count)) { atomic_read 492 net/bluetooth/bnep/core.c if (atomic_read(&s->terminate)) atomic_read 291 net/bluetooth/cmtp/core.c if (atomic_read(&session->terminate)) atomic_read 400 net/bluetooth/hci_conn.c int refcnt = atomic_read(&conn->refcnt); atomic_read 3594 net/bluetooth/hci_core.c if (atomic_read(&hdev->promisc)) { atomic_read 4460 net/bluetooth/hci_core.c if (atomic_read(&hdev->promisc)) { atomic_read 4517 net/bluetooth/hci_core.c atomic_read(&hdev->cmd_cnt), skb_queue_len(&hdev->cmd_q)); atomic_read 4520 net/bluetooth/hci_core.c if (atomic_read(&hdev->cmd_cnt)) { atomic_read 3485 net/bluetooth/hci_event.c if (atomic_read(&hdev->cmd_cnt) && !skb_queue_empty(&hdev->cmd_q)) atomic_read 3598 net/bluetooth/hci_event.c if (atomic_read(&hdev->cmd_cnt) && !skb_queue_empty(&hdev->cmd_q)) atomic_read 303 net/bluetooth/hci_sock.c if (!atomic_read(&monitor_promisc)) atomic_read 731 net/bluetooth/hci_sock.c if (atomic_read(&monitor_promisc)) { atomic_read 108 net/bluetooth/hidp/core.c if (atomic_read(&session->terminate)) atomic_read 244 net/bluetooth/hidp/core.c if (atomic_read(&session->terminate)) atomic_read 276 net/bluetooth/hidp/core.c !atomic_read(&session->terminate)) { atomic_read 281 net/bluetooth/hidp/core.c || atomic_read(&session->terminate), atomic_read 351 net/bluetooth/hidp/core.c !atomic_read(&session->terminate)) { atomic_read 356 net/bluetooth/hidp/core.c || atomic_read(&session->terminate), atomic_read 1065 net/bluetooth/hidp/core.c while (atomic_read(&session->state) <= HIDP_SESSION_IDLING) atomic_read 1067 net/bluetooth/hidp/core.c atomic_read(&session->state) > HIDP_SESSION_IDLING); atomic_read 1172 net/bluetooth/hidp/core.c atomic_read(&session->state) > HIDP_SESSION_PREPARING) atomic_read 1205 net/bluetooth/hidp/core.c if (atomic_read(&session->terminate)) atomic_read 1036 net/bluetooth/l2cap_sock.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf >> 1) atomic_read 1331 net/bluetooth/l2cap_sock.c lock_sock_nested(sk, atomic_read(&chan->nesting)); atomic_read 60 net/bluetooth/rfcomm/sock.c if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) atomic_read 643 net/bluetooth/rfcomm/sock.c if (atomic_read(&sk->sk_rmem_alloc) <= (sk->sk_rcvbuf >> 2)) atomic_read 356 net/bluetooth/rfcomm/tty.c int pending = 40 - atomic_read(&dev->wmem_alloc); atomic_read 449 net/bridge/br_multicast.c if (atomic_read(&br->mdb_hash_tbl.nelems) >= br->hash_max) { atomic_read 132 net/caif/caif_socket.c if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >= atomic_read 135 net/caif/caif_socket.c atomic_read(&cf_sk->sk.sk_rmem_alloc), atomic_read 261 net/caif/caif_socket.c if (atomic_read(&sk->sk_rmem_alloc) <= sk_rcvbuf_lowwater(cf_sk)) { atomic_read 134 net/caif/cfctrl.c req->sequence_no = atomic_read(&ctrl->req_seq_no); atomic_read 65 net/can/j1939/socket.c return atomic_read(&jsk->skb_pending); atomic_read 345 net/ceph/debugfs.c atomic_read(&osdc->num_requests), atomic_read 346 net/ceph/debugfs.c atomic_read(&osdc->num_homeless)); atomic_read 370 net/ceph/messenger.c if (atomic_read(&con->msgr->stopping)) { atomic_read 3343 net/ceph/osd_client.c if (atomic_read(&osdc->num_homeless) || !list_empty(&slow_osds)) atomic_read 5224 net/ceph/osd_client.c WARN_ON(atomic_read(&osdc->num_requests)); atomic_read 5225 net/ceph/osd_client.c WARN_ON(atomic_read(&osdc->num_homeless)); atomic_read 103 net/core/bpf_sk_storage.c atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) { atomic_read 1775 net/core/dev.c wanted = atomic_read(&netstamp_wanted); atomic_read 1795 net/core/dev.c wanted = atomic_read(&netstamp_wanted); atomic_read 1199 net/core/filter.c atomic_read(&sk->sk_omem_alloc) + filter_size < sysctl_optmem_max) { atomic_read 223 net/core/neighbour.c int max_clean = atomic_read(&tbl->gc_entries) - tbl->gc_thresh2; atomic_read 621 net/core/neighbour.c if (atomic_read(&tbl->entries) > (1 << nht->hash_shift)) atomic_read 911 net/core/neighbour.c if (atomic_read(&tbl->entries) < tbl->gc_thresh1) atomic_read 1076 net/core/neighbour.c atomic_read(&neigh->probes) >= neigh_max_probes(neigh)) { atomic_read 1733 net/core/neighbour.c if (atomic_read(&tbl->entries)) atomic_read 2058 net/core/neighbour.c .ndtc_entries = atomic_read(&tbl->entries), atomic_read 2463 net/core/neighbour.c if (nla_put_u32(skb, NDA_PROBES, atomic_read(&neigh->probes)) || atomic_read 3313 net/core/neighbour.c atomic_read(&tbl->entries), atomic_read 294 net/core/net-sysfs.c atomic_read(&netdev->carrier_up_count) + atomic_read 295 net/core/net-sysfs.c atomic_read(&netdev->carrier_down_count)); atomic_read 305 net/core/net-sysfs.c return sprintf(buf, fmt_dec, atomic_read(&netdev->carrier_up_count)); atomic_read 315 net/core/net-sysfs.c return sprintf(buf, fmt_dec, atomic_read(&netdev->carrier_down_count)); atomic_read 197 net/core/page_pool.c u32 release_cnt = atomic_read(&pool->pages_state_release_cnt); atomic_read 812 net/core/rtnetlink.c ci.rta_clntref = atomic_read(&dst->__refcnt); atomic_read 1635 net/core/rtnetlink.c atomic_read(&dev->carrier_up_count) + atomic_read 1636 net/core/rtnetlink.c atomic_read(&dev->carrier_down_count)) || atomic_read 1639 net/core/rtnetlink.c atomic_read(&dev->carrier_up_count)) || atomic_read 1641 net/core/rtnetlink.c atomic_read(&dev->carrier_down_count))) atomic_read 1155 net/core/skbuff.c next = (u32)atomic_read(&sk->sk_zckey); atomic_read 4417 net/core/skbuff.c if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >= atomic_read 703 net/core/skmsg.c if (atomic_read(&sk_other->sk_rmem_alloc) <= atomic_read 724 net/core/skmsg.c atomic_read(&sk_other->sk_rmem_alloc) <= atomic_read 456 net/core/sock.c if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) { atomic_read 1712 net/core/sock.c if (atomic_read(&sk->sk_omem_alloc)) atomic_read 1714 net/core/sock.c __func__, atomic_read(&sk->sk_omem_alloc)); atomic_read 2122 net/core/sock.c if (atomic_read(&sk->sk_omem_alloc) + SKB_TRUESIZE(size) > atomic_read 2142 net/core/sock.c atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) { atomic_read 2526 net/core/sock.c if (atomic_read(&sk->sk_rmem_alloc) < sk_get_rmem0(sk, prot)) atomic_read 2548 net/core/sock.c atomic_read(&sk->sk_rmem_alloc) + atomic_read 3219 net/core/sock.c mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc); atomic_read 3221 net/core/sock.c mem[SK_MEMINFO_DROPS] = atomic_read(&sk->sk_drops); atomic_read 344 net/dccp/proto.c if (atomic_read(&sk->sk_rmem_alloc) > 0) atomic_read 585 net/decnet/dn_nsp_in.c if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >= atomic_read 193 net/decnet/dn_route.c if (atomic_read(&rt->dst.__refcnt) > 1 || atomic_read 227 net/decnet/dn_route.c if (atomic_read(&rt->dst.__refcnt) > 1 || atomic_read 1838 net/decnet/dn_route.c atomic_read(&rt->dst.__refcnt), atomic_read 153 net/ipv4/af_inet.c WARN_ON(atomic_read(&sk->sk_rmem_alloc)); atomic_read 336 net/ipv4/arp.c int probes = atomic_read(&neigh->probes); atomic_read 1852 net/ipv4/devinet.c cb->seq = atomic_read(&tgt_net->ipv4.dev_addr_genid) ^ atomic_read 2252 net/ipv4/devinet.c cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^ atomic_read 1290 net/ipv4/fib_semantics.c nh->nh_saddr_genid = atomic_read(&net->ipv4.dev_addr_genid); atomic_read 1306 net/ipv4/fib_semantics.c if (nh->nh_saddr_genid == atomic_read(&net->ipv4.dev_addr_genid)) atomic_read 2180 net/ipv4/fib_semantics.c if (hash > atomic_read(&nexthop_nh->fib_nh_upper_bound)) atomic_read 505 net/ipv4/inet_connection_sock.c atomic_read(&newsk->sk_rmem_alloc)); atomic_read 161 net/ipv4/inet_timewait_sock.c if (atomic_read(&dr->tw_count) >= dr->sysctl_max_tw_buckets) atomic_read 1168 net/ipv4/ipmr.c if (atomic_read(&mrt->cache_resolve_queue_len) == 1) atomic_read 1331 net/ipv4/ipmr.c if (atomic_read(&mrt->cache_resolve_queue_len) != 0) { atomic_read 2756 net/ipv4/ipmr.c u32 queue_len = atomic_read(&mrt->cache_resolve_queue_len); atomic_read 504 net/ipv4/nexthop.c if (hash > atomic_read(&nhge->upper_bound)) atomic_read 1119 net/ipv4/ping.c atomic_read(&sp->sk_drops)); atomic_read 61 net/ipv4/proc.c atomic_read(&net->ipv4.tcp_death_row.tw_count), sockets, atomic_read 71 net/ipv4/proc.c atomic_read(&net->ipv4.fqdir->rhashtable.nelems), atomic_read 1082 net/ipv4/raw.c refcount_read(&sp->sk_refcnt), sp, atomic_read(&sp->sk_drops)); atomic_read 1557 net/ipv4/tcp.c !atomic_read(&sk->sk_rmem_alloc))) atomic_read 496 net/ipv4/tcp_fastopen.c int tfo_da_times = atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times); atomic_read 540 net/ipv4/tcp_fastopen.c atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times)) { atomic_read 488 net/ipv4/tcp_input.c min(atomic_read(&sk->sk_rmem_alloc), atomic_read 491 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf) atomic_read 4531 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf || atomic_read 5093 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf && atomic_read 5125 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) atomic_read 5130 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) atomic_read 5141 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) atomic_read 5149 net/ipv4/tcp_input.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) atomic_read 1457 net/ipv4/udp.c rmem = atomic_read(&sk->sk_rmem_alloc); atomic_read 2920 net/ipv4/udp.c atomic_read(&sp->sk_drops)); atomic_read 733 net/ipv6/addrconf.c cb->seq = atomic_read(&net->ipv6.dev_addr_genid) ^ atomic_read 5201 net/ipv6/addrconf.c cb->seq = atomic_read(&tgt_net->ipv6.dev_addr_genid) ^ tgt_net->dev_base_seq; atomic_read 1049 net/ipv6/datagram.c atomic_read(&sp->sk_drops)); atomic_read 95 net/ipv6/ip6_fib.c old = atomic_read(&net->ipv6.fib6_sernum); atomic_read 163 net/ipv6/ip6_flowlabel.c if (atomic_read(&fl->users) == 0) { atomic_read 180 net/ipv6/ip6_flowlabel.c if (!sched && atomic_read(&fl_size)) atomic_read 201 net/ipv6/ip6_flowlabel.c atomic_read(&fl->users) == 0) { atomic_read 469 net/ipv6/ip6_flowlabel.c int room = FL_MAX_SIZE - atomic_read(&fl_size); atomic_read 822 net/ipv6/ip6_flowlabel.c atomic_read(&fl->users), atomic_read 1529 net/ipv6/ip6mr.c if (atomic_read(&mrt->cache_resolve_queue_len) != 0) { atomic_read 721 net/ipv6/ndisc.c int probes = atomic_read(&neigh->probes); atomic_read 47 net/ipv6/proc.c atomic_read(&net->ipv6.fqdir->rhashtable.nelems), atomic_read 449 net/ipv6/route.c if (fl6->mp_hash <= atomic_read(&match->fib6_nh->fib_nh_upper_bound)) atomic_read 457 net/ipv6/route.c nh_upper_bound = atomic_read(&nh->fib_nh_upper_bound); atomic_read 6048 net/ipv6/route.c atomic_read(&net->ipv6.rt6_stats->fib_rt_alloc), atomic_read 295 net/iucv/af_iucv.c return ((atomic_read(&iucv->msg_sent) < iucv->msglimit_peer) && atomic_read 296 net/iucv/af_iucv.c (atomic_read(&iucv->pendings) <= 0)); atomic_read 335 net/iucv/af_iucv.c confirm_recv = atomic_read(&iucv->msg_recv); atomic_read 386 net/iucv/af_iucv.c WARN_ON(atomic_read(&iucv->msg_recv) < 0); atomic_read 418 net/iucv/af_iucv.c WARN_ON(atomic_read(&sk->sk_rmem_alloc)); atomic_read 1456 net/iucv/af_iucv.c if (atomic_read(&iucv->msg_recv) > iucv->msglimit) { atomic_read 1479 net/iucv/af_iucv.c if (atomic_read(&iucv->msg_recv) >= atomic_read 1860 net/iucv/af_iucv.c len = atomic_read(&sk->sk_rmem_alloc); atomic_read 2312 net/iucv/af_iucv.c if (atomic_read(&iucv->pendings) <= 0) atomic_read 150 net/kcm/kcmproc.c atomic_read(&psock->sk->sk_rmem_alloc), atomic_read 192 net/kcm/kcmsock.c if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) atomic_read 76 net/key/af_key.c if (3 * atomic_read(&sk->sk_rmem_alloc) <= 2 * sk->sk_rcvbuf) atomic_read 107 net/key/af_key.c WARN_ON(atomic_read(&sk->sk_rmem_alloc)); atomic_read 200 net/key/af_key.c if (atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf) atomic_read 3057 net/key/af_key.c if (atomic_read(&net_pfkey->socks_nr) == 0) atomic_read 3716 net/key/af_key.c 3 * atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) atomic_read 926 net/llc/llc_conn.c __func__, atomic_read(&llc_sock_nr)); atomic_read 978 net/llc/llc_conn.c atomic_read(&llc_sock_nr)); atomic_read 982 net/llc/llc_conn.c __func__, atomic_read(&llc_sock_nr)); atomic_read 2661 net/mac80211/cfg.c smps_mode, atomic_read(&sdata->u.ap.num_mcast_sta)); atomic_read 103 net/mac80211/debugfs_netdev.c return scnprintf(buf, buflen, "%d\n", atomic_read(&sdata->field));\ atomic_read 1584 net/mac80211/ieee80211_i.h return atomic_read(&sdata->u.ap.num_mcast_sta); atomic_read 1586 net/mac80211/ieee80211_i.h return atomic_read(&sdata->u.vlan.num_mcast_sta); atomic_read 14 net/mac80211/led.c if (!atomic_read(&local->assoc_led_active)) atomic_read 24 net/mac80211/led.c if (!atomic_read(&local->radio_led_active)) atomic_read 18 net/mac80211/led.h if (!atomic_read(&local->rx_led_active)) atomic_read 29 net/mac80211/led.h if (!atomic_read(&local->tx_led_active)) atomic_read 78 net/mac80211/led.h if (ieee80211_is_data(fc) && atomic_read(&local->tpt_led_active)) atomic_read 87 net/mac80211/led.h if (ieee80211_is_data(fc) && atomic_read(&local->tpt_led_active)) atomic_read 42 net/mac80211/main.c if (atomic_read(&local->iff_allmultis)) atomic_read 279 net/mac80211/mesh.c neighbors = atomic_read(&ifmsh->estab_plinks); atomic_read 338 net/mac80211/mesh.h atomic_read(&sdata->u.mesh.estab_plinks); atomic_read 440 net/mac80211/tx.c if (!atomic_read(&ps->num_sta_ps) && skb_queue_empty(&ps->bc_buf)) atomic_read 4219 net/mac80211/tx.c if (atomic_read(&ps->num_sta_ps) > 0) atomic_read 289 net/mac80211/util.c (ps && atomic_read(&ps->num_sta_ps)) || ac != vif->txq->ac) atomic_read 614 net/netfilter/ipset/ip_set_hash_gen.h if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) { atomic_read 1008 net/netfilter/ipset/ip_set_hash_gen.h if (atomic_read(&t->ref) && ext->target) { atomic_read 1035 net/netfilter/ipset/ip_set_hash_gen.h if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) { atomic_read 1096 net/netfilter/ipset/ip_set_hash_gen.h if (atomic_read(&t->ref) && ext->target) { atomic_read 1148 net/netfilter/ipset/ip_set_hash_gen.h if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) { atomic_read 1336 net/netfilter/ipset/ip_set_hash_gen.h if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) { atomic_read 585 net/netfilter/ipvs/ip_vs_app.c atomic_read(&inc->usecnt), atomic_read 298 net/netfilter/ipvs/ip_vs_conn.c if (!cp && atomic_read(&ip_vs_conn_no_cport_cnt)) { atomic_read 563 net/netfilter/ipvs/ip_vs_conn.c return atomic_read(&dest->activeconns) atomic_read 564 net/netfilter/ipvs/ip_vs_conn.c + atomic_read(&dest->inactconns); atomic_read 584 net/netfilter/ipvs/ip_vs_conn.c conn_flags = atomic_read(&dest->conn_flags); atomic_read 681 net/netfilter/ipvs/ip_vs_conn.c if (pd && atomic_read(&pd->appcnt)) atomic_read 744 net/netfilter/ipvs/ip_vs_conn.c (atomic_read(&dest->weight) == 0); atomic_read 818 net/netfilter/ipvs/ip_vs_conn.c if (atomic_read(&cp->n_control)) atomic_read 832 net/netfilter/ipvs/ip_vs_conn.c if (!cp->timeout && !atomic_read(&ct->n_control) && atomic_read 868 net/netfilter/ipvs/ip_vs_conn.c atomic_read(&cp->n_control)); atomic_read 984 net/netfilter/ipvs/ip_vs_conn.c if (unlikely(pd && atomic_read(&pd->appcnt))) atomic_read 1240 net/netfilter/ipvs/ip_vs_conn.c i = atomic_read(&cp->in_pkts); atomic_read 1276 net/netfilter/ipvs/ip_vs_conn.c if (atomic_read(&cp->n_control)) atomic_read 1364 net/netfilter/ipvs/ip_vs_conn.c if (atomic_read(&ipvs->conn_count) != 0) { atomic_read 1428 net/netfilter/ipvs/ip_vs_core.c if (atomic_read(&ipvs->conn_out_counter)) { atomic_read 2067 net/netfilter/ipvs/ip_vs_core.c unlikely(!atomic_read(&cp->dest->weight))) { atomic_read 2072 net/netfilter/ipvs/ip_vs_core.c if (!atomic_read(&cp->n_control)) { atomic_read 2084 net/netfilter/ipvs/ip_vs_core.c if (!atomic_read(&cp->n_control)) atomic_read 2363 net/netfilter/ipvs/ip_vs_core.c ipvs->gen = atomic_read(&ipvs_netns_cnt); atomic_read 225 net/netfilter/ipvs/ip_vs_ctl.c if (atomic_read(&ipvs->dropentry)) atomic_read 425 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&ipvs->ftpsvc_counter) && atomic_read 435 net/netfilter/ipvs/ip_vs_ctl.c && atomic_read(&ipvs->nullsvc_counter)) { atomic_read 2174 net/netfilter/ipvs/ip_vs_ctl.c ip_vs_fwd_name(atomic_read(&dest->conn_flags)), atomic_read 2175 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->weight), atomic_read 2176 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->activeconns), atomic_read 2177 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->inactconns)); atomic_read 2185 net/netfilter/ipvs/ip_vs_ctl.c ip_vs_fwd_name(atomic_read(&dest->conn_flags)), atomic_read 2186 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->weight), atomic_read 2187 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->activeconns), atomic_read 2188 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->inactconns)); atomic_read 2663 net/netfilter/ipvs/ip_vs_ctl.c entry.conn_flags = atomic_read(&dest->conn_flags); atomic_read 2664 net/netfilter/ipvs/ip_vs_ctl.c entry.weight = atomic_read(&dest->weight); atomic_read 2667 net/netfilter/ipvs/ip_vs_ctl.c entry.activeconns = atomic_read(&dest->activeconns); atomic_read 2668 net/netfilter/ipvs/ip_vs_ctl.c entry.inactconns = atomic_read(&dest->inactconns); atomic_read 2669 net/netfilter/ipvs/ip_vs_ctl.c entry.persistconns = atomic_read(&dest->persistconns); atomic_read 3257 net/netfilter/ipvs/ip_vs_ctl.c (atomic_read(&dest->conn_flags) & atomic_read 3260 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->weight)) || atomic_read 3270 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->activeconns)) || atomic_read 3272 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->inactconns)) || atomic_read 3274 net/netfilter/ipvs/ip_vs_ctl.c atomic_read(&dest->persistconns)) || atomic_read 225 net/netfilter/ipvs/ip_vs_dh.c || atomic_read(&dest->weight) <= 0 atomic_read 34 net/netfilter/ipvs/ip_vs_fo.c atomic_read(&dest->weight) > hw) { atomic_read 36 net/netfilter/ipvs/ip_vs_fo.c hw = atomic_read(&dest->weight); atomic_read 44 net/netfilter/ipvs/ip_vs_fo.c atomic_read(&hweight->activeconns), atomic_read 45 net/netfilter/ipvs/ip_vs_fo.c atomic_read(&hweight->weight)); atomic_read 311 net/netfilter/ipvs/ip_vs_lblc.c if (atomic_read(&tbl->entries) <= tbl->max_size) { atomic_read 316 net/netfilter/ipvs/ip_vs_lblc.c goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; atomic_read 420 net/netfilter/ipvs/ip_vs_lblc.c if (atomic_read(&dest->weight) > 0) { atomic_read 437 net/netfilter/ipvs/ip_vs_lblc.c if ((__s64)loh * atomic_read(&dest->weight) > atomic_read 438 net/netfilter/ipvs/ip_vs_lblc.c (__s64)doh * atomic_read(&least->weight)) { atomic_read 448 net/netfilter/ipvs/ip_vs_lblc.c atomic_read(&least->activeconns), atomic_read 450 net/netfilter/ipvs/ip_vs_lblc.c atomic_read(&least->weight), loh); atomic_read 463 net/netfilter/ipvs/ip_vs_lblc.c if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) { atomic_read 467 net/netfilter/ipvs/ip_vs_lblc.c if (atomic_read(&d->activeconns)*2 atomic_read 468 net/netfilter/ipvs/ip_vs_lblc.c < atomic_read(&d->weight)) { atomic_read 507 net/netfilter/ipvs/ip_vs_lblc.c atomic_read(&dest->weight) > 0 && !is_overloaded(dest, svc)) atomic_read 173 net/netfilter/ipvs/ip_vs_lblcr.c if ((atomic_read(&least->weight) > 0) atomic_read 189 net/netfilter/ipvs/ip_vs_lblcr.c if (((__s64)loh * atomic_read(&dest->weight) > atomic_read 190 net/netfilter/ipvs/ip_vs_lblcr.c (__s64)doh * atomic_read(&least->weight)) atomic_read 202 net/netfilter/ipvs/ip_vs_lblcr.c atomic_read(&least->activeconns), atomic_read 204 net/netfilter/ipvs/ip_vs_lblcr.c atomic_read(&least->weight), loh); atomic_read 222 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&most->weight) > 0) { atomic_read 235 net/netfilter/ipvs/ip_vs_lblcr.c if (((__s64)moh * atomic_read(&dest->weight) < atomic_read 236 net/netfilter/ipvs/ip_vs_lblcr.c (__s64)doh * atomic_read(&most->weight)) atomic_read 237 net/netfilter/ipvs/ip_vs_lblcr.c && (atomic_read(&dest->weight) > 0)) { atomic_read 247 net/netfilter/ipvs/ip_vs_lblcr.c atomic_read(&most->activeconns), atomic_read 249 net/netfilter/ipvs/ip_vs_lblcr.c atomic_read(&most->weight), moh); atomic_read 475 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&tbl->entries) <= tbl->max_size) { atomic_read 480 net/netfilter/ipvs/ip_vs_lblcr.c goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; atomic_read 584 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&dest->weight) > 0) { atomic_read 601 net/netfilter/ipvs/ip_vs_lblcr.c if ((__s64)loh * atomic_read(&dest->weight) > atomic_read 602 net/netfilter/ipvs/ip_vs_lblcr.c (__s64)doh * atomic_read(&least->weight)) { atomic_read 612 net/netfilter/ipvs/ip_vs_lblcr.c atomic_read(&least->activeconns), atomic_read 614 net/netfilter/ipvs/ip_vs_lblcr.c atomic_read(&least->weight), loh); atomic_read 627 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) { atomic_read 631 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&d->activeconns)*2 atomic_read 632 net/netfilter/ipvs/ip_vs_lblcr.c < atomic_read(&d->weight)) { atomic_read 663 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&en->set.size) > 1 && atomic_read 667 net/netfilter/ipvs/ip_vs_lblcr.c if (atomic_read(&en->set.size) > 1) { atomic_read 43 net/netfilter/ipvs/ip_vs_lc.c atomic_read(&dest->weight) == 0) atomic_read 59 net/netfilter/ipvs/ip_vs_lc.c atomic_read(&least->activeconns), atomic_read 60 net/netfilter/ipvs/ip_vs_lc.c atomic_read(&least->inactconns)); atomic_read 83 net/netfilter/ipvs/ip_vs_mh.c return atomic_read(&dest->weight) <= 0 || atomic_read 151 net/netfilter/ipvs/ip_vs_mh.c lw = atomic_read(&dest->last_weight); atomic_read 330 net/netfilter/ipvs/ip_vs_mh.c weight = atomic_read(&dest->last_weight); atomic_read 358 net/netfilter/ipvs/ip_vs_mh.c new_weight = atomic_read(&dest->last_weight); atomic_read 45 net/netfilter/ipvs/ip_vs_nq.c return atomic_read(&dest->activeconns) + 1; atomic_read 77 net/netfilter/ipvs/ip_vs_nq.c !atomic_read(&dest->weight)) atomic_read 83 net/netfilter/ipvs/ip_vs_nq.c if (atomic_read(&dest->activeconns) == 0) { atomic_read 90 net/netfilter/ipvs/ip_vs_nq.c ((__s64)loh * atomic_read(&dest->weight) > atomic_read 91 net/netfilter/ipvs/ip_vs_nq.c (__s64)doh * atomic_read(&least->weight))) { atomic_read 107 net/netfilter/ipvs/ip_vs_nq.c atomic_read(&least->activeconns), atomic_read 109 net/netfilter/ipvs/ip_vs_nq.c atomic_read(&least->weight), loh); atomic_read 36 net/netfilter/ipvs/ip_vs_ovf.c w = atomic_read(&dest->weight); atomic_read 38 net/netfilter/ipvs/ip_vs_ovf.c atomic_read(&dest->activeconns) > w || atomic_read 51 net/netfilter/ipvs/ip_vs_ovf.c atomic_read(&h->activeconns), atomic_read 52 net/netfilter/ipvs/ip_vs_ovf.c atomic_read(&h->weight)); atomic_read 71 net/netfilter/ipvs/ip_vs_rr.c atomic_read(&dest->weight) > 0) atomic_read 94 net/netfilter/ipvs/ip_vs_rr.c atomic_read(&dest->activeconns), atomic_read 95 net/netfilter/ipvs/ip_vs_rr.c refcount_read(&dest->refcnt), atomic_read(&dest->weight)); atomic_read 49 net/netfilter/ipvs/ip_vs_sed.c return atomic_read(&dest->activeconns) + 1; atomic_read 80 net/netfilter/ipvs/ip_vs_sed.c atomic_read(&dest->weight) > 0) { atomic_read 97 net/netfilter/ipvs/ip_vs_sed.c if ((__s64)loh * atomic_read(&dest->weight) > atomic_read 98 net/netfilter/ipvs/ip_vs_sed.c (__s64)doh * atomic_read(&least->weight)) { atomic_read 108 net/netfilter/ipvs/ip_vs_sed.c atomic_read(&least->activeconns), atomic_read 110 net/netfilter/ipvs/ip_vs_sed.c atomic_read(&least->weight), loh); atomic_read 76 net/netfilter/ipvs/ip_vs_sh.c return atomic_read(&dest->weight) <= 0 || atomic_read 192 net/netfilter/ipvs/ip_vs_sh.c atomic_read(&dest->weight)); atomic_read 195 net/netfilter/ipvs/ip_vs_sh.c if (++d_count >= atomic_read(&dest->weight)) { atomic_read 52 net/netfilter/ipvs/ip_vs_wlc.c atomic_read(&dest->weight) > 0) { atomic_read 69 net/netfilter/ipvs/ip_vs_wlc.c if ((__s64)loh * atomic_read(&dest->weight) > atomic_read 70 net/netfilter/ipvs/ip_vs_wlc.c (__s64)doh * atomic_read(&least->weight)) { atomic_read 80 net/netfilter/ipvs/ip_vs_wlc.c atomic_read(&least->activeconns), atomic_read 82 net/netfilter/ipvs/ip_vs_wlc.c atomic_read(&least->weight), loh); atomic_read 76 net/netfilter/ipvs/ip_vs_wrr.c weight = atomic_read(&dest->weight); atomic_read 97 net/netfilter/ipvs/ip_vs_wrr.c new_weight = atomic_read(&dest->weight); atomic_read 181 net/netfilter/ipvs/ip_vs_wrr.c atomic_read(&dest->weight) >= mark->cw) atomic_read 215 net/netfilter/ipvs/ip_vs_wrr.c atomic_read(&dest->activeconns), atomic_read 217 net/netfilter/ipvs/ip_vs_wrr.c atomic_read(&dest->weight)); atomic_read 342 net/netfilter/ipvs/ip_vs_xmit.c atomic_read(&rt->dst.__refcnt)); atomic_read 513 net/netfilter/ipvs/ip_vs_xmit.c atomic_read(&rt->dst.__refcnt)); atomic_read 601 net/netfilter/nf_conntrack_core.c WARN_ON(atomic_read(&nfct->use) != 0); atomic_read 1272 net/netfilter/nf_conntrack_core.c if (atomic_read(&net->ct.count) < nf_conntrack_max95) atomic_read 1357 net/netfilter/nf_conntrack_core.c unlikely(atomic_read(&net->ct.count) > nf_conntrack_max)) { atomic_read 1417 net/netfilter/nf_conntrack_core.c WARN_ON(atomic_read(&ct->ct_general.use) != 0); atomic_read 2164 net/netfilter/nf_conntrack_core.c if (atomic_read(&net->ct.count) > 0) { atomic_read 2180 net/netfilter/nf_conntrack_core.c if (atomic_read(&net->ct.count) == 0) atomic_read 2209 net/netfilter/nf_conntrack_core.c if (atomic_read(&net->ct.count) == 0) atomic_read 2290 net/netfilter/nf_conntrack_core.c if (atomic_read(&net->ct.count) != 0) atomic_read 501 net/netfilter/nf_conntrack_netlink.c if (nla_put_be32(skb, CTA_USE, htonl(atomic_read(&ct->ct_general.use)))) atomic_read 2289 net/netfilter/nf_conntrack_netlink.c unsigned int nr_conntracks = atomic_read(&net->ct.count); atomic_read 367 net/netfilter/nf_conntrack_standalone.c seq_printf(s, "use=%u\n", atomic_read(&ct->ct_general.use)); atomic_read 425 net/netfilter/nf_conntrack_standalone.c unsigned int nr_conntracks = atomic_read(&net->ct.count); atomic_read 239 net/netfilter/nf_nat_masquerade.c if (event != NETDEV_DOWN || atomic_read(&v6_worker_count) >= 16) atomic_read 3821 net/netfilter/nf_tables_api.c (nlh->nlmsg_flags & NLM_F_NONREC && atomic_read(&set->nelems) > 0)) { atomic_read 263 net/netfilter/nft_ct.c if (likely(atomic_read(&ct->ct_general.use) == 1)) { atomic_read 30 net/netfilter/nft_numgen.c oval = atomic_read(&priv->counter); atomic_read 42 net/netfilter/xt_statistic.c oval = atomic_read(&info->master->count); atomic_read 957 net/netlabel/netlabel_kapi.c return (atomic_read(&netlabel_mgmt_protocount) > 0); atomic_read 401 net/netlink/af_netlink.c WARN_ON(atomic_read(&sk->sk_rmem_alloc)); atomic_read 427 net/netlink/af_netlink.c if (atomic_read(&nl_table_users)) { atomic_read 433 net/netlink/af_netlink.c if (atomic_read(&nl_table_users) == 0) atomic_read 1210 net/netlink/af_netlink.c if ((atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf || atomic_read 1224 net/netlink/af_netlink.c if ((atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf || atomic_read 1379 net/netlink/af_netlink.c if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf && atomic_read 1383 net/netlink/af_netlink.c return atomic_read(&sk->sk_rmem_alloc) > (sk->sk_rcvbuf >> 1); atomic_read 2000 net/netlink/af_netlink.c atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf / 2) { atomic_read 2206 net/netlink/af_netlink.c if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) atomic_read 2637 net/netlink/af_netlink.c atomic_read(&s->sk_drops), atomic_read 419 net/netlink/genetlink.c atomic_read(&genl_sk_destructing_cnt) == 0); atomic_read 136 net/netrom/nr_timer.c if (atomic_read(&sk->sk_rmem_alloc) < (sk->sk_rcvbuf / 2) && atomic_read 463 net/nfc/llcp_core.c pr_debug("%d clients\n", atomic_read(client_cnt)); atomic_read 786 net/nfc/nci/core.c if ((atomic_read(&ndev->state) == NCI_DISCOVERY) || atomic_read 787 net/nfc/nci/core.c (atomic_read(&ndev->state) == NCI_W4_ALL_DISCOVERIES)) { atomic_read 797 net/nfc/nci/core.c if ((atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) || atomic_read 798 net/nfc/nci/core.c (atomic_read(&ndev->state) == NCI_POLL_ACTIVE)) { atomic_read 837 net/nfc/nci/core.c if ((atomic_read(&ndev->state) != NCI_DISCOVERY) && atomic_read 838 net/nfc/nci/core.c (atomic_read(&ndev->state) != NCI_W4_ALL_DISCOVERIES)) { atomic_read 858 net/nfc/nci/core.c if ((atomic_read(&ndev->state) != NCI_W4_HOST_SELECT) && atomic_read 859 net/nfc/nci/core.c (atomic_read(&ndev->state) != NCI_POLL_ACTIVE)) { atomic_read 887 net/nfc/nci/core.c if (atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) { atomic_read 935 net/nfc/nci/core.c if (atomic_read(&ndev->state) == NCI_POLL_ACTIVE) { atomic_read 972 net/nfc/nci/core.c if (atomic_read(&ndev->state) == NCI_LISTEN_ACTIVE || atomic_read 973 net/nfc/nci/core.c atomic_read(&ndev->state) == NCI_DISCOVERY) { atomic_read 1432 net/nfc/nci/core.c pr_debug("credits_cnt %d\n", atomic_read(&conn_info->credits_cnt)); atomic_read 1435 net/nfc/nci/core.c while (atomic_read(&conn_info->credits_cnt)) { atomic_read 1441 net/nfc/nci/core.c if (atomic_read(&conn_info->credits_cnt) != atomic_read 1510 net/nfc/nci/core.c pr_debug("cmd_cnt %d\n", atomic_read(&ndev->cmd_cnt)); atomic_read 1513 net/nfc/nci/core.c if (atomic_read(&ndev->cmd_cnt)) { atomic_read 72 net/nfc/nci/ntf.c if (atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) { atomic_read 637 net/nfc/nci/ntf.c if (atomic_read(&ndev->state) == NCI_DISCOVERY) { atomic_read 180 net/nfc/nci/rsp.c (atomic_read(&ndev->state) != NCI_POLL_ACTIVE)) { atomic_read 661 net/packet/af_packet.c while (atomic_read(&pkc->blk_fill_in_prog)) { atomic_read 762 net/packet/af_packet.c if (atomic_read(&po->tp_drops)) atomic_read 923 net/packet/af_packet.c while (atomic_read(&pkc->blk_fill_in_prog)) { atomic_read 1236 net/packet/af_packet.c int avail = rcvbuf - atomic_read(&sk->sk_rmem_alloc) atomic_read 1286 net/packet/af_packet.c WARN_ON(atomic_read(&sk->sk_rmem_alloc)); atomic_read 2090 net/packet/af_packet.c if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) atomic_read 2242 net/packet/af_packet.c atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf) { atomic_read 2299 net/packet/af_packet.c if (atomic_read(&po->tp_drops)) atomic_read 4304 net/packet/af_packet.c if (atomic_read(&po->mapped)) atomic_read 4407 net/packet/af_packet.c if (closing || atomic_read(&po->mapped) == 0) { atomic_read 4425 net/packet/af_packet.c if (atomic_read(&po->mapped)) atomic_read 4427 net/packet/af_packet.c atomic_read(&po->mapped)); atomic_read 4609 net/packet/af_packet.c atomic_read(&s->sk_rmem_alloc), atomic_read 1158 net/phonet/pep.c done = atomic_read(&pn->tx_credits); atomic_read 1172 net/phonet/pep.c done = sk_wait_event(sk, &timeo, atomic_read(&pn->tx_credits), &wait); atomic_read 1194 net/phonet/pep.c return atomic_read(&pn->tx_credits); atomic_read 350 net/phonet/socket.c atomic_read(&pn->tx_credits)) atomic_read 604 net/phonet/socket.c atomic_read(&sk->sk_drops)); atomic_read 284 net/rds/cong.c unsigned long gen = atomic_read(&rds_cong_generation); atomic_read 375 net/rds/connection.c atomic_read(&cp->cp_state)); atomic_read 407 net/rds/connection.c atomic_read(&cp->cp_state)); atomic_read 748 net/rds/connection.c atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING, atomic_read 751 net/rds/connection.c atomic_read(&cp->cp_state) == RDS_CONN_UP, atomic_read 774 net/rds/connection.c atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING, atomic_read 777 net/rds/connection.c atomic_read(&cp->cp_state) == RDS_CONN_UP, atomic_read 322 net/rds/ib.c iinfo->cache_allocs = atomic_read(&ic->i_cache_allocs); atomic_read 359 net/rds/ib.c iinfo6->cache_allocs = atomic_read(&ic->i_cache_allocs); atomic_read 505 net/rds/ib.c atomic_read(&rds_ib_unloading) != 0); atomic_read 255 net/rds/ib_cm.c (atomic_read(&ic->i_credits)); atomic_read 323 net/rds/ib_cm.c if (atomic_read(&ic->i_cq_quiesce)) atomic_read 368 net/rds/ib_cm.c if (atomic_read(&ic->i_cq_quiesce)) atomic_read 1021 net/rds/ib_cm.c (atomic_read(&ic->i_signaled_sends) == 0) && atomic_read 1022 net/rds/ib_cm.c (atomic_read(&ic->i_fastreg_inuse_count) == 0) && atomic_read 1023 net/rds/ib_cm.c (atomic_read(&ic->i_fastreg_wrs) == RDS_IB_DEFAULT_FR_WR)); atomic_read 47 net/rds/ib_fmr.c if (atomic_read(&pool->dirty_count) >= pool->max_items / 10) atomic_read 51 net/rds/ib_fmr.c if (atomic_read(&pool->dirty_count) >= pool->max_items * 9 / 10) { atomic_read 92 net/rds/ib_frmr.c if (atomic_read(&pool->item_count) > pool->max_items_soft) atomic_read 118 net/rds/ib_frmr.c if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned || atomic_read 119 net/rds/ib_frmr.c atomic_read(&pool->dirty_count) >= pool->max_items / 5) atomic_read 275 net/rds/ib_rdma.c item_count = atomic_read(&pool->item_count); atomic_read 495 net/rds/ib_rdma.c if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned || atomic_read 496 net/rds/ib_rdma.c atomic_read(&pool->dirty_count) >= pool->max_items / 5) atomic_read 574 net/rds/ib_rdma.c WARN_ON(atomic_read(&pool->item_count)); atomic_read 575 net/rds/ib_rdma.c WARN_ON(atomic_read(&pool->free_pinned)); atomic_read 1076 net/rds/ib_recv.c WARN_ON(atomic_read(&rds_ib_allocation)); atomic_read 78 net/rds/ib_ring.c diff = ring->w_alloc_ctr - (u32) atomic_read(&ring->w_free_ctr); atomic_read 232 net/rds/ib_send.c BUG_ON(atomic_read(&ic->i_signaled_sends) < 0); atomic_read 366 net/rds/ib_send.c oldval = newval = atomic_read(&ic->i_credits); atomic_read 416 net/rds/ib_send.c IB_GET_SEND_CREDITS(atomic_read(&ic->i_credits)), atomic_read 449 net/rds/ib_send.c if (IB_GET_POST_CREDITS(atomic_read(&ic->i_credits)) >= 16) atomic_read 55 net/rds/loop.c return atomic_read(&rds_loop_unloading) != 0; atomic_read 814 net/rds/rds.h return atomic_read(&cp->cp_state); atomic_read 827 net/rds/rds.h return atomic_read(&cp->cp_state) == RDS_CONN_UP; atomic_read 840 net/rds/rds.h return atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING; atomic_read 434 net/rds/tcp.c return atomic_read(&rds_tcp_unloading) != 0; atomic_read 80 net/rds/threads.c atomic_read(&cp->cp_state)); atomic_read 418 net/rfkill/core.c if (atomic_read(&rfkill_input_disabled)) atomic_read 443 net/rfkill/core.c if (atomic_read(&rfkill_input_disabled)) atomic_read 471 net/rfkill/core.c if (atomic_read(&rfkill_input_disabled)) atomic_read 490 net/rfkill/core.c if (atomic_read(&rfkill_input_disabled)) atomic_read 1051 net/rfkill/core.c if (!atomic_read(&rfkill_input_disabled)) atomic_read 181 net/rose/rose_in.c if (atomic_read(&sk->sk_rmem_alloc) > atomic_read 141 net/rose/rose_timer.c if (atomic_read(&sk->sk_rmem_alloc) < (sk->sk_rcvbuf / 2) && atomic_read 353 net/rxrpc/af_rxrpc.c _enter("%d{%d}", call->debug_id, atomic_read(&call->usage)); atomic_read 1047 net/rxrpc/af_rxrpc.c ASSERTCMP(atomic_read(&rxrpc_n_tx_skbs), ==, 0); atomic_read 1048 net/rxrpc/af_rxrpc.c ASSERTCMP(atomic_read(&rxrpc_n_rx_skbs), ==, 0); atomic_read 19 net/rxrpc/ar-internal.h BUG_ON(atomic_read((X)) >> (sizeof(atomic_t) - 2) == \ atomic_read 88 net/rxrpc/call_accept.c atomic_read(&conn->usage), here); atomic_read 101 net/rxrpc/call_accept.c atomic_read(&call->usage), atomic_read 91 net/rxrpc/call_object.c _leave(" = %p [%d]", call, atomic_read(&call->usage)); atomic_read 243 net/rxrpc/call_object.c atomic_read(&call->usage), atomic_read 294 net/rxrpc/call_object.c atomic_read(&call->usage), here, NULL); atomic_read 317 net/rxrpc/call_object.c atomic_read(&call->usage), here, ERR_PTR(ret)); atomic_read 392 net/rxrpc/call_object.c int n = atomic_read(&call->usage); atomic_read 409 net/rxrpc/call_object.c int n = atomic_read(&call->usage); atomic_read 449 net/rxrpc/call_object.c _enter("{%d,%d}", call->debug_id, atomic_read(&call->usage)); atomic_read 452 net/rxrpc/call_object.c atomic_read(&call->usage), atomic_read 639 net/rxrpc/call_object.c call, atomic_read(&call->usage), atomic_read 652 net/rxrpc/call_object.c wait_var_event(&rxnet->nr_calls, !atomic_read(&rxnet->nr_calls)); atomic_read 156 net/rxrpc/conn_client.c conn, atomic_read(&conn->usage)); atomic_read 216 net/rxrpc/conn_client.c atomic_read(&conn->usage), atomic_read 110 net/rxrpc/conn_object.c if (!conn || atomic_read(&conn->usage) == 0) atomic_read 120 net/rxrpc/conn_object.c if (!conn || atomic_read(&conn->usage) == 0) { atomic_read 284 net/rxrpc/conn_object.c int n = atomic_read(&conn->usage); atomic_read 354 net/rxrpc/conn_object.c _enter("{%d,u=%d}", conn->debug_id, atomic_read(&conn->usage)); atomic_read 356 net/rxrpc/conn_object.c ASSERTCMP(atomic_read(&conn->usage), ==, 0); atomic_read 395 net/rxrpc/conn_object.c ASSERTCMP(atomic_read(&conn->usage), >, 0); atomic_read 396 net/rxrpc/conn_object.c if (likely(atomic_read(&conn->usage) > 1)) atomic_read 408 net/rxrpc/conn_object.c conn->debug_id, atomic_read(&conn->usage), atomic_read 445 net/rxrpc/conn_object.c ASSERTCMP(atomic_read(&conn->usage), ==, 0); atomic_read 473 net/rxrpc/conn_object.c conn, atomic_read(&conn->usage)); atomic_read 484 net/rxrpc/conn_object.c wait_var_event(&rxnet->nr_conns, !atomic_read(&rxnet->nr_conns)); atomic_read 96 net/rxrpc/conn_service.c if (atomic_read(&cursor->usage) == 0) atomic_read 138 net/rxrpc/conn_service.c atomic_read(&conn->usage), atomic_read 1403 net/rxrpc/input.c if (!call || atomic_read(&call->usage) == 0) { atomic_read 344 net/rxrpc/local_object.c int n = atomic_read(&local->usage); atomic_read 452 net/rxrpc/local_object.c atomic_read(&local->usage), NULL); atomic_read 508 net/rxrpc/local_object.c local, atomic_read(&local->usage)); atomic_read 124 net/rxrpc/peer_object.c atomic_read(&peer->usage) > 0) atomic_read 143 net/rxrpc/peer_object.c _leave(" = %p {u=%d}", peer, atomic_read(&peer->usage)); atomic_read 373 net/rxrpc/peer_object.c _leave(" = %p {u=%d}", peer, atomic_read(&peer->usage)); atomic_read 478 net/rxrpc/peer_object.c atomic_read(&peer->usage), atomic_read 110 net/rxrpc/proc.c atomic_read(&call->usage), atomic_read 192 net/rxrpc/proc.c atomic_read(&conn->usage), atomic_read 195 net/rxrpc/proc.c atomic_read(&conn->serial), atomic_read 242 net/rxrpc/proc.c atomic_read(&peer->usage), atomic_read 38 net/rxrpc/skbuff.c int n = atomic_read(select_skb_count(skb)); atomic_read 162 net/sched/act_api.c if (!bind && strict && atomic_read(&p->tcfa_bindcnt) > 0) atomic_read 275 net/sched/act_api.c if (atomic_read(&p->tcfa_bindcnt) > 0) atomic_read 379 net/sched/act_api.c if (!atomic_read(&p->tcfa_bindcnt)) { atomic_read 142 net/sched/act_bpf.c .bindcnt = atomic_read(&prog->tcf_bindcnt) - bind, atomic_read 177 net/sched/act_connmark.c .bindcnt = atomic_read(&ci->tcf_bindcnt) - bind, atomic_read 641 net/sched/act_csum.c .bindcnt = atomic_read(&p->tcf_bindcnt) - bind, atomic_read 839 net/sched/act_ct.c .bindcnt = atomic_read(&c->tcf_bindcnt) - bind, atomic_read 289 net/sched/act_ctinfo.c .bindcnt = atomic_read(&ci->tcf_bindcnt) - bind, atomic_read 200 net/sched/act_gact.c .bindcnt = atomic_read(&gact->tcf_bindcnt) - bind, atomic_read 627 net/sched/act_ife.c .bindcnt = atomic_read(&ife->tcf_bindcnt) - bind, atomic_read 294 net/sched/act_ipt.c c.bindcnt = atomic_read(&ipt->tcf_bindcnt) - bind; atomic_read 336 net/sched/act_mirred.c .bindcnt = atomic_read(&m->tcf_bindcnt) - bind, atomic_read 306 net/sched/act_mpls.c .bindcnt = atomic_read(&m->tcf_bindcnt) - bind, atomic_read 270 net/sched/act_nat.c .bindcnt = atomic_read(&p->tcf_bindcnt) - bind, atomic_read 435 net/sched/act_pedit.c opt->bindcnt = atomic_read(&p->tcf_bindcnt) - bind; atomic_read 313 net/sched/act_police.c .bindcnt = atomic_read(&police->tcf_bindcnt) - bind, atomic_read 207 net/sched/act_sample.c .bindcnt = atomic_read(&s->tcf_bindcnt) - bind, atomic_read 178 net/sched/act_simple.c .bindcnt = atomic_read(&d->tcf_bindcnt) - bind, atomic_read 237 net/sched/act_skbedit.c .bindcnt = atomic_read(&d->tcf_bindcnt) - bind, atomic_read 223 net/sched/act_skbmod.c .bindcnt = atomic_read(&d->tcf_bindcnt) - bind, atomic_read 521 net/sched/act_tunnel_key.c .bindcnt = atomic_read(&t->tcf_bindcnt) - bind, atomic_read 261 net/sched/act_vlan.c .bindcnt = atomic_read(&v->tcf_bindcnt) - bind, atomic_read 702 net/sched/cls_api.c return atomic_read(&block->offloadcnt); atomic_read 416 net/sched/em_meta.c dst->value = atomic_read(&sk->sk_omem_alloc); atomic_read 413 net/sctp/associola.c WARN_ON(atomic_read(&asoc->rmem_alloc)); atomic_read 1436 net/sctp/associola.c atomic_read(&t->mtu_info)); atomic_read 1538 net/sctp/associola.c rx_count = atomic_read(&asoc->rmem_alloc); atomic_read 1540 net/sctp/associola.c rx_count = atomic_read(&asoc->base.sk->sk_rmem_alloc); atomic_read 169 net/sctp/diag.c amt = atomic_read(&asoc->rmem_alloc); atomic_read 177 net/sctp/diag.c mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc); atomic_read 179 net/sctp/diag.c mem[SK_MEMINFO_DROPS] = atomic_read(&sk->sk_drops); atomic_read 421 net/sctp/diag.c r->idiag_rqueue = atomic_read(&infox->asoc->rmem_alloc); atomic_read 69 net/sctp/objcnt.c atomic_read(sctp_dbg_objcnt[i].counter)); atomic_read 259 net/sctp/proc.c atomic_read(&assoc->rmem_alloc), atomic_read 9432 net/sctp/socket.c if (atomic_read(&sctp_sk(oldsk)->pd_mode)) { atomic_read 633 net/sctp/ulpevent.c rx_count = atomic_read(&asoc->rmem_alloc); atomic_read 635 net/sctp/ulpevent.c rx_count = atomic_read(&sk->sk_rmem_alloc); atomic_read 215 net/sctp/ulpqueue.c if (atomic_read(&sp->pd_mode) == 0) { atomic_read 483 net/sctp/ulpqueue.c atomic_read(&sctp_sk(asoc->base.sk)->pd_mode)) atomic_read 1059 net/sctp/ulpqueue.c if (sp->frag_interleave || atomic_read(&sp->pd_mode) == 0) { atomic_read 1494 net/smc/af_smc.c } else if (!atomic_read(&smc_sk(nsk)->conn.bytes_to_rcv)) { atomic_read 1633 net/smc/af_smc.c atomic_read(&smc->conn.sndbuf_space)) || atomic_read 1640 net/smc/af_smc.c if (atomic_read(&smc->conn.bytes_to_rcv)) atomic_read 1817 net/smc/af_smc.c answ = atomic_read(&smc->conn.bytes_to_rcv); atomic_read 1830 net/smc/af_smc.c atomic_read(&smc->conn.sndbuf_space); atomic_read 287 net/smc/smc_cdc.c atomic_read(&conn->peer_rmbe_space) == conn->peer_rmbe_size) { atomic_read 94 net/smc/smc_close.c if (atomic_read(&conn->bytes_to_rcv)) atomic_read 187 net/smc/smc_rx.c return atomic_read(&conn->bytes_to_rcv) && atomic_read 188 net/smc/smc_rx.c !atomic_read(&conn->splice_pending); atomic_read 367 net/smc/smc_rx.c readable = atomic_read(&conn->bytes_to_rcv); atomic_read 368 net/smc/smc_rx.c splbytes = atomic_read(&conn->splice_pending); atomic_read 28 net/smc/smc_rx.h return atomic_read(&conn->bytes_to_rcv); atomic_read 47 net/smc/smc_tx.c if (atomic_read(&smc->conn.sndbuf_space) && sock) { atomic_read 108 net/smc/smc_tx.c if (atomic_read(&conn->sndbuf_space) && !conn->urg_tx_pend) atomic_read 115 net/smc/smc_tx.c (atomic_read(&conn->sndbuf_space) && atomic_read 166 net/smc/smc_tx.c if (!atomic_read(&conn->sndbuf_space) || conn->urg_tx_pend) { atomic_read 177 net/smc/smc_tx.c writespace = atomic_read(&conn->sndbuf_space); atomic_read 225 net/smc/smc_tx.c (atomic_read(&conn->sndbuf_space) > atomic_read 421 net/smc/smc_tx.c rmbespace = atomic_read(&conn->peer_rmbe_space); atomic_read 583 net/smc/smc_tx.c !atomic_read(&conn->bytes_to_rcv)) atomic_read 619 net/smc/smc_tx.c !atomic_read(&conn->bytes_to_rcv)) atomic_read 605 net/sunrpc/auth_gss/gss_krb5_mech.c if (seq_send64 != atomic_read(&ctx->seq_send)) { atomic_read 607 net/sunrpc/auth_gss/gss_krb5_mech.c seq_send64, atomic_read(&ctx->seq_send)); atomic_read 249 net/sunrpc/backchannel_rqst.c if (atomic_read(&xprt->bc_slot_count) >= BC_MAX_SLOTS) atomic_read 1172 net/sunrpc/cache.c if (atomic_read(&detail->writers)) atomic_read 173 net/sunrpc/clnt.c if (atomic_read(&clnt->cl_count) == 0) atomic_read 1083 net/sunrpc/clnt.c if (atomic_read(&clnt->cl_swapper)) atomic_read 1226 net/sunrpc/clnt.c WARN_ON_ONCE(atomic_read(&task->tk_count) != 2); atomic_read 1211 net/sunrpc/sched.c if (atomic_read(&task->tk_count) != 1 + !RPC_IS_ASYNC(task)) { atomic_read 1613 net/sunrpc/svc.c WARN_ON_ONCE(atomic_read(&task->tk_count) != 1); atomic_read 344 net/sunrpc/svc_xprt.c int nrqsts = atomic_read(&xprt->xpt_nr_rqsts); atomic_read 629 net/sunrpc/svcsock.c required = atomic_read(&svsk->sk_xprt.xpt_reserved) + serv->sv_max_mesg; atomic_read 986 net/sunrpc/xprt.c return atomic_read(&req->rq_pin) != 0; atomic_read 95 net/sunrpc/xprtrdma/svc_rdma.c int len = snprintf(str_buf, 32, "%d\n", atomic_read(stat)); atomic_read 334 net/sunrpc/xprtrdma/svc_rdma_rw.c atomic_read(&rdma->sc_sq_avail) > cc->cc_sqecount); atomic_read 301 net/sunrpc/xprtrdma/svc_rdma_sendto.c atomic_read(&rdma->sc_sq_avail) > 1); atomic_read 1282 net/sunrpc/xprtsock.c if (atomic_read(&transport->xprt.swapper)) atomic_read 2067 net/sunrpc/xprtsock.c if (atomic_read(&xprt->swapper)) atomic_read 2268 net/tipc/socket.c lim = rcvbuf_limit(sk, skb) + atomic_read(dcnt); atomic_read 3473 net/tipc/socket.c atomic_read(&sk->sk_drops))) atomic_read 3718 net/tipc/socket.c unsigned int lim = rcvbuf_limit(sk, skb) + atomic_read(dcnt); atomic_read 3794 net/tipc/socket.c i += scnprintf(buf + i, sz - i, " %u", atomic_read(&tsk->dupl_rcvcnt)); atomic_read 371 net/tipc/topsrv.c if (atomic_read(&tn->subscription_count) >= TIPC_MAX_SUBSCR) { atomic_read 1103 net/tls/tls_sw.c pending = atomic_read(&ctx->encrypt_pending); atomic_read 1906 net/tls/tls_sw.c pending = atomic_read(&ctx->decrypt_pending); atomic_read 2127 net/tls/tls_sw.c if (atomic_read(&ctx->encrypt_pending)) atomic_read 368 net/vmw_vsock/virtio_transport.c val = atomic_read(&vsock->queued_replies); atomic_read 96 net/wireless/lib80211.c if (atomic_read(&entry->refcnt) != 0 && !force) atomic_read 81 net/wireless/scan.c if (WARN_ON(atomic_read(&bss->hold))) atomic_read 404 net/wireless/scan.c if (atomic_read(&bss->hold)) atomic_read 425 net/wireless/scan.c if (atomic_read(&bss->hold)) atomic_read 945 net/wireless/scan.c !atomic_read(&bss->hold)) atomic_read 2071 net/wireless/scan.c WARN_ON(atomic_read(&new->hold)); atomic_read 291 net/x25/x25_in.c if (atomic_read(&sk->sk_rmem_alloc) > atomic_read 370 net/x25/x25_subr.c if (atomic_read(&sk->sk_rmem_alloc) < (sk->sk_rcvbuf >> 1) && atomic_read 2744 net/xfrm/xfrm_policy.c xdst->policy_genid = atomic_read(&pols[0]->genid); atomic_read 3809 net/xfrm/xfrm_policy.c xdst->policy_genid != atomic_read(&xdst->pols[0]->genid)) atomic_read 2401 net/xfrm/xfrm_state.c if (atomic_read(&t->tunnel_users) == 2) atomic_read 118 security/integrity/ima/ima_main.c if (atomic_read(&inode->i_readcount) && IS_IMA(inode)) { atomic_read 156 security/integrity/ima/ima_main.c if (atomic_read(&inode->i_writecount) == 1) { atomic_read 315 security/keys/proc.c atomic_read(&user->nkeys), atomic_read 316 security/keys/proc.c atomic_read(&user->nikeys), atomic_read 176 security/selinux/avc.c atomic_read(&avc->avc_cache.active_nodes), atomic_read 165 security/selinux/hooks.c atomic_read(&selinux_secmark_refcount)); atomic_read 36 security/selinux/include/xfrm.h return (atomic_read(&selinux_xfrm_refcount) > 0); atomic_read 2083 security/tomoyo/common.c if (atomic_read(&tomoyo_query_observers)) atomic_read 2127 security/tomoyo/common.c !atomic_read(&tomoyo_query_observers), HZ)) atomic_read 2360 security/tomoyo/common.c atomic_read(&tomoyo_stat_updated[i])); atomic_read 420 security/tomoyo/condition.c atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS) atomic_read 451 security/tomoyo/gc.c if (atomic_read(&container_of atomic_read 538 security/tomoyo/gc.c if (!domain->is_deleted || atomic_read(&domain->users)) atomic_read 555 security/tomoyo/gc.c if (atomic_read(&ptr->users) > 0) atomic_read 581 security/tomoyo/gc.c atomic_read(&group->head.users) > 0) atomic_read 596 security/tomoyo/gc.c if (atomic_read(&ptr->users) > 0) atomic_read 114 security/tomoyo/memory.c atomic_read(&group->head.users) == TOMOYO_GC_IN_PROGRESS) atomic_read 168 security/tomoyo/memory.c atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS) atomic_read 697 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) { atomic_read 816 sound/core/oss/pcm_oss.c if (atomic_read(&runtime->oss.rw_ref)) { atomic_read 853 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 862 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 989 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count) || atomic_read 998 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count) || atomic_read 1374 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 1484 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 1639 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 1818 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 2057 sound/core/oss/pcm_oss.c if (atomic_read(&psubstream->mmap_count)) atomic_read 2185 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) { atomic_read 2771 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 2782 sound/core/oss/pcm_oss.c if (atomic_read(&substream->mmap_count)) atomic_read 658 sound/core/pcm_native.c if (atomic_read(&substream->mmap_count)) atomic_read 781 sound/core/pcm_native.c if (atomic_read(&substream->mmap_count)) atomic_read 423 sound/core/seq/seq_clientmgr.c if (atomic_read(&fifo->overflow) > 0) { atomic_read 686 sound/core/seq/seq_clientmgr.c if (atomic_read(&subs->ref_count) != 2) atomic_read 1863 sound/core/seq/seq_clientmgr.c if (atomic_read(&client->pool->counter)) atomic_read 16 sound/core/seq/seq_lock.c if (atomic_read(lockp) < 0) { atomic_read 17 sound/core/seq/seq_lock.c pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line); atomic_read 20 sound/core/seq/seq_lock.c while (atomic_read(lockp) > 0) { atomic_read 22 sound/core/seq/seq_lock.c pr_warn("ALSA: seq_lock: waiting [%d left] in %s:%d\n", atomic_read(lockp), file, line); atomic_read 24 sound/core/seq/seq_memory.c return pool->total_elements - atomic_read(&pool->counter); atomic_read 257 sound/core/seq/seq_memory.c used = atomic_read(&pool->counter); atomic_read 432 sound/core/seq/seq_memory.c while (atomic_read(&pool->counter) > 0) atomic_read 498 sound/core/seq/seq_memory.c snd_iprintf(buffer, "%sCells in use : %d\n", space, atomic_read(&pool->counter)); atomic_read 60 sound/core/seq/seq_memory.h return pool ? pool->total_elements - atomic_read(&pool->counter) : 0; atomic_read 372 sound/drivers/dummy.c if (!atomic_read(&dpcm->running)) atomic_read 379 sound/drivers/dummy.c if (!atomic_read(&dpcm->running)) atomic_read 48 sound/drivers/pcsp/pcsp_input.c if (atomic_read(&pcsp_chip.timer_active) || !pcsp_chip.pcspkr) atomic_read 31 sound/drivers/pcsp/pcsp_lib.c if (atomic_read(&pcsp_chip.timer_active)) { atomic_read 135 sound/drivers/pcsp/pcsp_lib.c if (!atomic_read(&chip->timer_active) || !chip->playback_substream) atomic_read 158 sound/drivers/pcsp/pcsp_lib.c if (atomic_read(&chip->timer_active)) { atomic_read 179 sound/drivers/pcsp/pcsp_lib.c if (!atomic_read(&chip->timer_active)) atomic_read 318 sound/drivers/pcsp/pcsp_lib.c if (atomic_read(&chip->timer_active)) { atomic_read 718 sound/isa/gus/gus_pcm.c if (!wait_event_timeout(pcmp->sleep, (atomic_read(&pcmp->dma_count) <= 0), 2*HZ)) atomic_read 357 sound/pci/echoaudio/echoaudio.c if (atomic_read(&chip->opencount) > 1 && chip->rate_set) atomic_read 360 sound/pci/echoaudio/echoaudio.c chip->can_set_rate, atomic_read(&chip->opencount), atomic_read 392 sound/pci/echoaudio/echoaudio.c if (atomic_read(&chip->opencount) > 1 && chip->rate_set) atomic_read 395 sound/pci/echoaudio/echoaudio.c chip->can_set_rate, atomic_read(&chip->opencount), atomic_read 433 sound/pci/echoaudio/echoaudio.c if (atomic_read(&chip->opencount) > 1 && chip->rate_set) atomic_read 475 sound/pci/echoaudio/echoaudio.c if (atomic_read(&chip->opencount) > 1 && chip->rate_set) atomic_read 498 sound/pci/echoaudio/echoaudio.c oc = atomic_read(&chip->opencount); atomic_read 1447 sound/pci/echoaudio/echoaudio.c if (atomic_read(&chip->opencount)) { atomic_read 305 sound/pci/echoaudio/mona_dsp.c if (atomic_read(&chip->opencount)) atomic_read 464 sound/pci/mixart/mixart.c while (atomic_read(&mgr->msg_processed) > 0) { atomic_read 193 sound/pci/ymfpci/ymfpci_main.c if (atomic_read(&chip->interrupt_sleep_count)) { atomic_read 816 sound/pci/ymfpci/ymfpci_main.c if (atomic_read(&chip->interrupt_sleep_count)) { atomic_read 1625 sound/soc/codecs/hdac_hdmi.c if (atomic_read(&hdev->in_pm)) atomic_read 51 sound/soc/fsl/imx-pcm-fiq.c if (!atomic_read(&iprtd->playing) && !atomic_read(&iprtd->capturing)) atomic_read 132 sound/soc/fsl/imx-pcm-fiq.c if (!atomic_read(&iprtd->playing) && atomic_read 133 sound/soc/fsl/imx-pcm-fiq.c !atomic_read(&iprtd->capturing)) atomic_read 38 sound/soc/intel/atom/sst/sst_drv_interface.c #define GET_USAGE_COUNT(dev) (atomic_read(&dev->power.usage_count)) atomic_read 345 sound/usb/card.c if (!atomic_read(&chip->shutdown)) atomic_read 589 sound/usb/card.c if (atomic_read(&usb_chip[i]->shutdown)) { atomic_read 717 sound/usb/card.c !atomic_read(&chip->usage_count)); atomic_read 761 sound/usb/card.c if (atomic_read(&chip->shutdown)) { atomic_read 788 sound/usb/card.c if (atomic_read(&chip->shutdown)) atomic_read 797 sound/usb/card.c if (atomic_read(&chip->shutdown)) atomic_read 369 sound/usb/endpoint.c if (unlikely(atomic_read(&ep->chip->shutdown))) atomic_read 552 sound/usb/endpoint.c if (!force && atomic_read(&ep->chip->shutdown)) /* to be sure... */ atomic_read 926 sound/usb/endpoint.c if (atomic_read(&ep->chip->shutdown)) atomic_read 371 sound/usb/midi.c if (atomic_read(&in->urbs[j]->use_count)) atomic_read 71 sound/usb/pcm.c if (atomic_read(&subs->stream->chip->shutdown)) atomic_read 36 sound/usb/proc.c if (!atomic_read(&chip->shutdown)) atomic_read 43 sound/usb/proc.c if (!atomic_read(&chip->shutdown)) atomic_read 126 sound/usb/usx2y/us122l.c snd_printdd(KERN_DEBUG "%i\n", atomic_read(&us122l->mmap_count)); atomic_read 169 sound/usb/usx2y/us122l.c snd_printdd(KERN_DEBUG "%i\n", atomic_read(&us122l->mmap_count)); atomic_read 639 sound/usb/usx2y/us122l.c while (atomic_read(&us122l->mmap_count)) atomic_read 137 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&subs->state) >= state_PRERUNNING) atomic_read 205 sound/usb/usx2y/usbusx2yaudio.c state = atomic_read(&playbacksubs->state); atomic_read 232 sound/usb/usx2y/usbusx2yaudio.c state = atomic_read(&capsubs->state); atomic_read 254 sound/usb/usx2y/usbusx2yaudio.c snd_printdd("%i %p state=%i\n", s, subs, atomic_read(&subs->state)); atomic_read 261 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&subs->state) >= state_PRERUNNING) atomic_read 288 sound/usb/usx2y/usbusx2yaudio.c if (unlikely(atomic_read(&subs->state) < state_PREPARED)) { atomic_read 306 sound/usb/usx2y/usbusx2yaudio.c atomic_read(&capsubs->state) >= state_PREPARED && atomic_read 308 sound/usb/usx2y/usbusx2yaudio.c atomic_read(&playbacksubs->state) < state_PREPARED)) { atomic_read 462 sound/usb/usx2y/usbusx2yaudio.c if (subs != NULL && atomic_read(&subs->state) >= state_PREPARED) atomic_read 495 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&subs->state) != state_PREPARED) atomic_read 524 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&subs->state) == state_PREPARED && atomic_read 525 sound/usb/usx2y/usbusx2yaudio.c atomic_read(&subs->usX2Y->subs[SNDRV_PCM_STREAM_CAPTURE]->state) >= state_PREPARED) { atomic_read 534 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&subs->state) >= state_PRERUNNING) atomic_read 805 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&playback_subs->state) < state_PREPARED) { atomic_read 831 sound/usb/usx2y/usbusx2yaudio.c if (atomic_read(&capsubs->state) < state_PREPARED) { atomic_read 843 sound/usb/usx2y/usbusx2yaudio.c if (subs != capsubs && atomic_read(&subs->state) < state_PREPARED) atomic_read 125 sound/usb/usx2y/usx2yhwdeppcm.c if (atomic_read(&subs->state) != state_RUNNING) atomic_read 168 sound/usb/usx2y/usx2yhwdeppcm.c state = atomic_read(&playbacksubs->state); atomic_read 195 sound/usb/usx2y/usx2yhwdeppcm.c state = atomic_read(&capsubs->state); atomic_read 224 sound/usb/usx2y/usx2yhwdeppcm.c if (unlikely(atomic_read(&subs->state) < state_PREPARED)) { atomic_read 240 sound/usb/usx2y/usx2yhwdeppcm.c if (capsubs->completed_urb && atomic_read(&capsubs->state) >= state_PREPARED && atomic_read 242 sound/usb/usx2y/usx2yhwdeppcm.c (playbacksubs->completed_urb || atomic_read(&playbacksubs->state) < state_PREPARED)) { atomic_read 369 sound/usb/usx2y/usx2yhwdeppcm.c if (atomic_read(&playback_subs->state) < state_PREPARED) { atomic_read 413 sound/usb/usx2y/usx2yhwdeppcm.c if (subs != NULL && atomic_read(&subs->state) >= state_PREPARED) atomic_read 453 sound/usb/usx2y/usx2yhwdeppcm.c if (atomic_read(&subs->state) != state_PREPARED) atomic_read 490 sound/usb/usx2y/usx2yhwdeppcm.c if (atomic_read(&capsubs->state) < state_PREPARED) { atomic_read 505 sound/usb/usx2y/usx2yhwdeppcm.c if (atomic_read(&subs->state) < state_PREPARED) { atomic_read 587 sound/xen/xen_snd_front_alsa.c new_hw_ptr = (snd_pcm_uframes_t)atomic_read(&stream->hw_ptr); atomic_read 602 sound/xen/xen_snd_front_alsa.c return (snd_pcm_uframes_t)atomic_read(&stream->hw_ptr); atomic_read 65 tools/include/linux/refcount.h return atomic_read(&r->refs); atomic_read 78 tools/include/linux/refcount.h unsigned int old, new, val = atomic_read(&r->refs); atomic_read 123 tools/include/linux/refcount.h unsigned int old, new, val = atomic_read(&r->refs); atomic_read 906 virt/kvm/arm/arm.c int nrcpus = atomic_read(&kvm->online_vcpus); atomic_read 58 virt/kvm/arm/vgic/vgic-debug.c int nr_cpus = atomic_read(&kvm->online_vcpus); atomic_read 112 virt/kvm/arm/vgic/vgic-init.c if (atomic_read(&kvm->online_vcpus) > kvm->arch.max_vcpus) { atomic_read 273 virt/kvm/arm/vgic/vgic-init.c if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus)) atomic_read 1216 virt/kvm/arm/vgic/vgic-its.c if (target_addr >= atomic_read(&kvm->online_vcpus)) atomic_read 1348 virt/kvm/arm/vgic/vgic-its.c if (target1_addr >= atomic_read(&kvm->online_vcpus) || atomic_read 1349 virt/kvm/arm/vgic/vgic-its.c target2_addr >= atomic_read(&kvm->online_vcpus)) atomic_read 1836 virt/kvm/arm/vgic/vgic-its.c sz = atomic_read(&kvm->online_vcpus) * LPI_DEFAULT_PCPU_CACHE_SIZE; atomic_read 2476 virt/kvm/arm/vgic/vgic-its.c target_addr >= atomic_read(&kvm->online_vcpus)) atomic_read 289 virt/kvm/arm/vgic/vgic-kvm-device.c if (cpuid >= atomic_read(&dev->kvm->online_vcpus)) atomic_read 311 virt/kvm/arm/vgic/vgic-kvm-device.c unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1); atomic_read 38 virt/kvm/arm/vgic/vgic-mmio-v2.c value |= (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5; atomic_read 112 virt/kvm/arm/vgic/vgic-mmio-v2.c int nr_vcpus = atomic_read(&source_vcpu->kvm->online_vcpus); atomic_read 174 virt/kvm/arm/vgic/vgic-mmio-v2.c u8 cpu_mask = GENMASK(atomic_read(&vcpu->kvm->online_vcpus) - 1, 0); atomic_read 116 virt/kvm/arm/vgic/vgic-v4.c nr_vcpus = atomic_read(&kvm->online_vcpus); atomic_read 289 virt/kvm/arm/vgic/vgic.h return atomic_read(&kvm->online_vcpus) * KVM_VGIC_V3_REDIST_SIZE; atomic_read 2827 virt/kvm/kvm_main.c BUG_ON(kvm->vcpus[atomic_read(&kvm->online_vcpus)]); atomic_read 2837 virt/kvm/kvm_main.c kvm->vcpus[atomic_read(&kvm->online_vcpus)] = vcpu; atomic_read 3757 virt/kvm/kvm_main.c if (atomic_read(&hardware_enable_failed)) {