atomic_read       654 arch/alpha/kernel/smp.c 		if (atomic_read(&mm->mm_users) <= 1) {
atomic_read       701 arch/alpha/kernel/smp.c 		if (atomic_read(&mm->mm_users) <= 1) {
atomic_read       755 arch/alpha/kernel/smp.c 		if (atomic_read(&mm->mm_users) <= 1) {
atomic_read       301 arch/arc/mm/tlb.c 	if (atomic_read(&mm->mm_users) == 0)
atomic_read       188 arch/arm/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       112 arch/arm/kernel/machine_kexec.c 	while ((atomic_read(&waiting_for_crash_ipi) > 0) && msecs) {
atomic_read       116 arch/arm/kernel/machine_kexec.c 	if (atomic_read(&waiting_for_crash_ipi) > 0)
atomic_read       155 arch/arm/kvm/coproc.c 	ncores = atomic_read(&vcpu->kvm->online_vcpus) - 1;
atomic_read       172 arch/arm/kvm/coproc.c 	if (atomic_read(&vcpu->kvm->online_vcpus) > 1)
atomic_read       210 arch/arm/mach-exynos/pm.c 			if (atomic_read(&cpu1_wakeup))
atomic_read       253 arch/arm/mach-exynos/pm.c 			       !atomic_read(&cpu1_wakeup))
atomic_read       256 arch/arm/mach-exynos/pm.c 			if (!atomic_read(&cpu1_wakeup))
atomic_read       260 arch/arm/mach-exynos/pm.c 		while (!atomic_read(&cpu1_wakeup)) {
atomic_read       255 arch/arm64/include/asm/pgtable.h 	if (mm != current->active_mm && atomic_read(&mm->mm_users) <= 1)
atomic_read       232 arch/arm64/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       204 arch/arm64/kernel/insn.c 		while (atomic_read(&pp->cpu_count) <= num_online_cpus())
atomic_read      1031 arch/arm64/kernel/smp.c 	while ((atomic_read(&waiting_for_crash_ipi) > 0) && timeout--)
atomic_read      1034 arch/arm64/kernel/smp.c 	if (atomic_read(&waiting_for_crash_ipi) > 0)
atomic_read      1043 arch/arm64/kernel/smp.c 	return (atomic_read(&waiting_for_crash_ipi) > 0);
atomic_read       145 arch/csky/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       157 arch/hexagon/kernel/kgdb.c 	if (atomic_read(&kgdb_active) != -1) {
atomic_read        40 arch/ia64/include/asm/atomic.h 		old = atomic_read(v);					\
atomic_read        55 arch/ia64/include/asm/atomic.h 		old = atomic_read(v);					\
atomic_read        85 arch/ia64/include/asm/tlbflush.h 	if (atomic_read(&mm->mm_users) == 0)
atomic_read        70 arch/ia64/kernel/crash.c 		if (atomic_read(&kdump_cpu_frozen) == cpu_num)
atomic_read       161 arch/ia64/kernel/crash.c 	if (atomic_read(&kdump_in_progress)) {
atomic_read        49 arch/ia64/kernel/irq.c 	seq_printf(p, "ERR: %10u\n", atomic_read(&irq_err_count));
atomic_read       383 arch/ia64/kernel/kprobes.c 	i = atomic_read(&kcb->prev_kprobe_index);
atomic_read      1604 arch/ia64/kernel/mca.c 	if (atomic_read(&kdump_in_progress))
atomic_read      1707 arch/ia64/kernel/mca.c 		while (monarch_cpu == -1 && !atomic_read(&kdump_in_progress))
atomic_read      1718 arch/ia64/kernel/mca.c 		while (monarch_cpu != -1 && !atomic_read(&kdump_in_progress))
atomic_read       298 arch/ia64/kernel/smp.c 	if (likely(mm == current->active_mm && atomic_read(&mm->mm_users) == 1))
atomic_read       124 arch/ia64/kernel/uncached.c 		if (atomic_read(&uc_pool->status))
atomic_read       143 arch/ia64/kernel/uncached.c 	if (atomic_read(&uc_pool->status))
atomic_read       408 arch/ia64/pci/pci.c 	BUG_ON(atomic_read(&dev->enable_cnt));
atomic_read       118 arch/m68k/amiga/chipram.c 	unsigned long n = atomic_read(&chipavail);
atomic_read        53 arch/m68k/include/asm/atomic.h 			: "g" (i), "2" (atomic_read(v)));		\
atomic_read        68 arch/m68k/include/asm/atomic.h 			: "g" (i), "2" (atomic_read(v)));		\
atomic_read       176 arch/m68k/include/asm/atomic.h 	prev = atomic_read(v);
atomic_read       189 arch/m68k/include/asm/atomic.h 	prev = atomic_read(v);
atomic_read        37 arch/m68k/kernel/irq.c 	seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count));
atomic_read        32 arch/microblaze/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read        94 arch/mips/cavium-octeon/setup.c 	while (!atomic_read(&kexec_ready_to_reboot))
atomic_read        47 arch/mips/kernel/crash.c 	while (!atomic_read(&kexec_ready_to_reboot))
atomic_read       332 arch/mips/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read        43 arch/mips/kernel/irq.c 	seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count));
atomic_read       324 arch/mips/kernel/kgdb.c 	if (atomic_read(&kgdb_active) != -1)
atomic_read       332 arch/mips/kernel/kgdb.c 	if (atomic_read(&kgdb_setting_breakpoint))
atomic_read       126 arch/mips/kernel/machine_kexec.c 	while (!atomic_read(&kexec_ready_to_reboot))
atomic_read       632 arch/mips/kernel/perf_event_mipsxx.c 		if (atomic_read(&active_events) == 0)
atomic_read        98 arch/mips/kernel/pm-cps.c 	while (atomic_read(a) < online)
atomic_read       106 arch/mips/kernel/pm-cps.c 	while (atomic_read(a) > online)
atomic_read       538 arch/mips/kernel/smp.c 	} else if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
atomic_read       588 arch/mips/kernel/smp.c 	} else if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
atomic_read       655 arch/mips/kernel/smp.c 	} else if ((atomic_read(&vma->vm_mm->mm_users) != 1) ||
atomic_read        67 arch/mips/kernel/spinlock_test.c 	while (atomic_read(&s->enter_wait))
atomic_read        74 arch/mips/kernel/spinlock_test.c 	while (atomic_read(&s->start_wait))
atomic_read        86 arch/mips/kernel/spinlock_test.c 	while (atomic_read(&s->exit_wait))
atomic_read        50 arch/mips/kernel/sync-r4k.c 		while (atomic_read(&count_count_start) != 1)
atomic_read        71 arch/mips/kernel/sync-r4k.c 		while (atomic_read(&count_count_stop) != 1)
atomic_read       101 arch/mips/kernel/sync-r4k.c 		while (atomic_read(&count_count_start) != 2)
atomic_read       111 arch/mips/kernel/sync-r4k.c 		while (atomic_read(&count_count_stop) != 2)
atomic_read       214 arch/mips/kernel/traps.c 		} else if (atomic_read(&kgdb_active) != -1 &&
atomic_read       164 arch/mips/kvm/mips.c 	for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
atomic_read       183 arch/mips/math-emu/dsemul.c 	fr_idx = atomic_read(&current->thread.bd_emu_frame);
atomic_read       249 arch/mips/math-emu/dsemul.c 	fr_idx = atomic_read(&current->thread.bd_emu_frame);
atomic_read       237 arch/mips/sgi-ip27/ip27-nmi.c 	while (atomic_read(&nmied_cpus) != num_online_cpus());
atomic_read       215 arch/nds32/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       871 arch/nds32/kernel/perf_event_cpu.c 		if (atomic_read(active_events) == 0) {
atomic_read        53 arch/openrisc/kernel/sync-timer.c 		while (atomic_read(&count_count_start) != 1)
atomic_read        74 arch/openrisc/kernel/sync-timer.c 		while (atomic_read(&count_count_stop) != 1)
atomic_read       104 arch/openrisc/kernel/sync-timer.c 		while (atomic_read(&count_count_start) != 2)
atomic_read       114 arch/openrisc/kernel/sync-timer.c 		while (atomic_read(&count_count_stop) != 2)
atomic_read        25 arch/parisc/include/asm/mmu_context.h 	BUG_ON(atomic_read(&mm->mm_users) != 1);
atomic_read        40 arch/parisc/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       146 arch/powerpc/include/asm/book3s/64/tlbflush.h 	if (atomic_read(&vma->vm_mm->context.copros) > 0)
atomic_read       114 arch/powerpc/include/asm/eeh.h 	return pe ? !!atomic_read(&pe->pass_dev_cnt) : false;
atomic_read       526 arch/powerpc/include/asm/kvm_book3s_64.h 	if (atomic_read(&kvm->arch.hpte_mod_interest))
atomic_read        15 arch/powerpc/include/asm/membarrier.h 	if (likely(!(atomic_read(&next->membarrier_state) &
atomic_read        66 arch/powerpc/include/asm/tlb.h 	if (atomic_read(&mm->context.active_cpus) > 1)
atomic_read        72 arch/powerpc/include/asm/tlb.h 	WARN_ON(atomic_read(&mm->context.copros) > 0);
atomic_read       126 arch/powerpc/kernel/crash.c 	while ((atomic_read(&cpus_in_crash) < ncpus) && (--msecs > 0))
atomic_read       131 arch/powerpc/kernel/crash.c 	if (atomic_read(&cpus_in_crash) >= ncpus) {
atomic_read       137 arch/powerpc/kernel/crash.c 		ncpus - atomic_read(&cpus_in_crash));
atomic_read       168 arch/powerpc/kernel/crash.c 		while (atomic_read(&cpus_in_crash) < ncpus)
atomic_read       754 arch/powerpc/kernel/rtas.c 	while (rc == H_MULTI_THREADS_ACTIVE && !atomic_read(&data->done) &&
atomic_read       755 arch/powerpc/kernel/rtas.c 	       !atomic_read(&data->error))
atomic_read       758 arch/powerpc/kernel/rtas.c 	if (rc || atomic_read(&data->error)) {
atomic_read       763 arch/powerpc/kernel/rtas.c 	if (atomic_read(&data->error))
atomic_read       764 arch/powerpc/kernel/rtas.c 		rc = atomic_read(&data->error);
atomic_read       800 arch/powerpc/kernel/rtas.c 	while (rc == H_SUCCESS && !atomic_read(&data->done) && !atomic_read(&data->error))
atomic_read      1000 arch/powerpc/kernel/rtas.c 	if (atomic_read(&data.error) != 0)
atomic_read      1015 arch/powerpc/kernel/rtas.c 	return atomic_read(&data.error);
atomic_read       383 arch/powerpc/kernel/smp.c 		spin_until_cond(atomic_read(&__nmi_ipi_lock) == 0);
atomic_read       392 arch/powerpc/kernel/smp.c 		spin_until_cond(atomic_read(&__nmi_ipi_lock) == 0);
atomic_read       398 arch/powerpc/kernel/smp.c 	WARN_ON(atomic_read(&__nmi_ipi_lock) != 1);
atomic_read       602 arch/powerpc/kernel/time.c 	if (atomic_read(&ppc_n_lost_interrupts) != 0)
atomic_read       955 arch/powerpc/kernel/trace/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       136 arch/powerpc/kvm/book3s_64_mmu_hv.c 		if (atomic_read(&kvm->arch.vcpus_running)) {
atomic_read      1041 arch/powerpc/kvm/book3s_64_mmu_hv.c 	return atomic_read(&kvm->arch.vcpus_running) != 0;
atomic_read      1866 arch/powerpc/kvm/book3s_64_mmu_hv.c 		if (atomic_read(&kvm->arch.vcpus_running)) {
atomic_read      3230 arch/powerpc/kvm/book3s_hv.c 		int n_online = atomic_read(&vc->online_count);
atomic_read      5323 arch/powerpc/kvm/book3s_hv.c 			if (atomic_read(&kvm->arch.vcpus_running)) {
atomic_read       184 arch/powerpc/kvm/book3s_hv_builtin.c 	return atomic_read(&hv_vm_count) != 0;
atomic_read       200 arch/powerpc/kvm/book3s_xive_template.c 			if (atomic_read(&q->pending_count)) {
atomic_read       204 arch/powerpc/kvm/book3s_xive_template.c 					WARN_ON(p > atomic_read(&q->count));
atomic_read       479 arch/powerpc/kvm/powerpc.c 	for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
atomic_read       344 arch/powerpc/mm/book3s64/pgtable.c 	BUG_ON(atomic_read(&page->pt_frag_refcount) <= 0);
atomic_read       106 arch/powerpc/mm/book3s64/radix_hugetlbpage.c 	    (atomic_read(&mm->context.copros) > 0))
atomic_read      1044 arch/powerpc/mm/book3s64/radix_pgtable.c 	if ((change & _PAGE_RW) && atomic_read(&mm->context.copros) > 0) {
atomic_read      1078 arch/powerpc/mm/book3s64/radix_pgtable.c 	    (atomic_read(&mm->context.copros) > 0))
atomic_read       364 arch/powerpc/mm/book3s64/radix_tlb.c 	if (atomic_read(&mm->context.copros) > 0)
atomic_read       497 arch/powerpc/mm/book3s64/radix_tlb.c 	if (atomic_read(&mm->context.copros) > 0)
atomic_read       551 arch/powerpc/mm/book3s64/radix_tlb.c 	if (atomic_read(&mm->context.copros) > 0)
atomic_read       617 arch/powerpc/mm/book3s64/radix_tlb.c 	if (atomic_read(&mm->context.copros) > 0)
atomic_read       619 arch/powerpc/mm/book3s64/radix_tlb.c 	if (atomic_read(&mm->mm_users) <= 1 && current->mm == mm)
atomic_read       631 arch/powerpc/mm/book3s64/radix_tlb.c 	if (atomic_read(&mm->context.copros) > 0)
atomic_read       279 arch/powerpc/mm/hugetlbpage.c 	if (atomic_read(&tlb->mm->mm_users) < 2 ||
atomic_read       113 arch/powerpc/mm/pgtable-frag.c 	BUG_ON(atomic_read(&page->pt_frag_refcount) <= 0);
atomic_read        41 arch/powerpc/perf/8xx-pmu.c 		ctr = atomic_read(&instruction_counter);
atomic_read        43 arch/powerpc/perf/8xx-pmu.c 	} while (ctr != atomic_read(&instruction_counter));
atomic_read      1990 arch/powerpc/perf/core-book3s.c 		if (atomic_read(&num_events) == 0 &&
atomic_read       222 arch/powerpc/perf/core-fsl-emb.c 		if (atomic_read(&num_events)) {
atomic_read       571 arch/powerpc/perf/core-fsl-emb.c 		if (atomic_read(&num_events) == 0 &&
atomic_read       441 arch/powerpc/platforms/85xx/smp.c 	while ( (atomic_read(&kexec_down_cpus) != (num_cpus - 1)) &&
atomic_read        37 arch/powerpc/platforms/cell/cpufreq_spudemand.c 	busy_spus = atomic_read(&cbe_spu_info[cpu_to_node(cpu)].busy_spus);
atomic_read      1459 arch/powerpc/platforms/cell/spufs/file.c 	if (atomic_read(&inode->i_count) != 1)
atomic_read       354 arch/powerpc/platforms/cell/spufs/inode.c 			if ((cbe_spu_info[node].n_spus - atomic_read(
atomic_read       405 arch/powerpc/platforms/cell/spufs/sched.c 	if (atomic_read(&ctx->gang->aff_sched_count) == 0)
atomic_read      1079 arch/powerpc/platforms/cell/spufs/sched.c 		atomic_read(&nr_spu_contexts),
atomic_read       103 arch/powerpc/platforms/powermac/backlight.c 	if (atomic_read(&kernel_backlight_disabled))
atomic_read       131 arch/powerpc/platforms/powermac/backlight.c 	if (atomic_read(&kernel_backlight_disabled))
atomic_read       170 arch/powerpc/platforms/powermac/backlight.c 	if (atomic_read(&kernel_backlight_disabled))
atomic_read       178 arch/powerpc/platforms/powermac/backlight.c 	if (atomic_read(&kernel_backlight_disabled))
atomic_read       645 arch/powerpc/platforms/powernv/idle.c 		if (atomic_read(&local_paca->dont_stop)) {
atomic_read      1230 arch/powerpc/platforms/powernv/vas-window.c 	if (!window->tx_win && atomic_read(&window->num_txwins) != 0) {
atomic_read        64 arch/powerpc/platforms/pseries/suspend.c 	if (atomic_read(&suspending))
atomic_read       944 arch/powerpc/platforms/pseries/vio.c 	return sprintf(buf, "%d\n", atomic_read(&viodev->cmo.allocs_failed));
atomic_read       168 arch/powerpc/sysdev/xive/common.c 		if (atomic_read(&q->pending_count)) {
atomic_read       171 arch/powerpc/sysdev/xive/common.c 				WARN_ON(p > atomic_read(&q->count));
atomic_read       882 arch/powerpc/xmon/xmon.c 		if (!bp->enabled && atomic_read(&bp->ref_count) == 0) {
atomic_read       136 arch/riscv/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       116 arch/s390/include/asm/mmu_context.h 		while (atomic_read(&mm->context.flush_count))
atomic_read       209 arch/s390/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       277 arch/s390/kernel/perf_cpum_cf.c 		if (atomic_read(&num_events) == 0 && __kernel_cpumcf_begin())
atomic_read       452 arch/s390/kernel/perf_cpum_cf.c 	if (!atomic_read(&cpuhw->ctr_set[event->hw.config_base]))
atomic_read       188 arch/s390/kernel/perf_cpum_cf_diag.c 			    atomic_read(&cf_diag_events));
atomic_read       249 arch/s390/kernel/perf_cpum_cf_diag.c 			    attr->sample_type, atomic_read(&cf_diag_events));
atomic_read       778 arch/s390/kernel/perf_cpum_sf.c 		if (atomic_read(&num_events) == 0 && reserve_pmc_hardware())
atomic_read      2087 arch/s390/kernel/perf_cpum_sf.c 	if (!atomic_read(&num_events))
atomic_read       372 arch/s390/kernel/time.c 	sw0 = atomic_read(sw_ptr);
atomic_read       374 arch/s390/kernel/time.c 	sw1 = atomic_read(sw_ptr);
atomic_read       422 arch/s390/kernel/time.c 	rc = (atomic_read(sw_ptr) & 0x80000000U) != 0;
atomic_read       604 arch/s390/kernel/time.c 		while (atomic_read(&sync->cpus) != 0)
atomic_read       361 arch/s390/kernel/topology.c 	if (atomic_read(&topology_poll) > 60)
atomic_read      1814 arch/s390/kvm/interrupt.c 	online_vcpus = atomic_read(&kvm->online_vcpus);
atomic_read      2987 arch/s390/kvm/interrupt.c 	int vcpu_id, online_vcpus = atomic_read(&kvm->online_vcpus);
atomic_read      1919 arch/s390/kvm/kvm-s390.c 	int slot = atomic_read(&slots->lru_slot);
atomic_read      2553 arch/s390/kvm/kvm-s390.c 	for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
atomic_read      3122 arch/s390/kvm/kvm-s390.c 	return atomic_read(&vcpu->arch.sie_block->prog20) &
atomic_read      3732 arch/s390/kvm/kvm-s390.c 	cpuflags = atomic_read(&vcpu->arch.sie_block->cpuflags);
atomic_read      4165 arch/s390/kvm/kvm-s390.c 	online_vcpus = atomic_read(&vcpu->kvm->online_vcpus);
atomic_read      4205 arch/s390/kvm/kvm-s390.c 	online_vcpus = atomic_read(&vcpu->kvm->online_vcpus);
atomic_read        60 arch/s390/kvm/kvm-s390.h 	return (atomic_read(&vcpu->arch.sie_block->cpuflags) & flags) == flags;
atomic_read       811 arch/s390/kvm/priv.c 	cpus = atomic_read(&vcpu->kvm->online_vcpus);
atomic_read        87 arch/s390/kvm/vsie.c 	return !(atomic_read(&vsie_page->scb_s.prog20) & PROG_REQUEST);
atomic_read        96 arch/s390/kvm/vsie.c 	cpuflags = atomic_read(&vsie_page->scb_o->cpuflags);
atomic_read       106 arch/s390/kvm/vsie.c 	int newflags, cpuflags = atomic_read(&scb_o->cpuflags);
atomic_read       486 arch/s390/kvm/vsie.c 	if (!(atomic_read(&scb_s->cpuflags) & CPUSTAT_KSS))
atomic_read       491 arch/s390/kvm/vsie.c 	if (!(atomic_read(&scb_s->cpuflags) & CPUSTAT_SM))
atomic_read      1068 arch/s390/kvm/vsie.c 		if (!(atomic_read(&scb_o->cpuflags) & CPUSTAT_STOP_INT))
atomic_read      1236 arch/s390/kvm/vsie.c 	nr_vcpus = atomic_read(&kvm->online_vcpus);
atomic_read       160 arch/s390/mm/pgalloc.c 		old = atomic_read(v);
atomic_read       205 arch/s390/mm/pgalloc.c 			mask = atomic_read(&page->_refcount) >> 24;
atomic_read       175 arch/sh/drivers/dma/dma-api.c 		if (atomic_read(&channel->busy) == 0)
atomic_read       102 arch/sh/kernel/ftrace.c 	int old = atomic_read(&nmi_running);
atomic_read       159 arch/sh/kernel/ftrace.c 	if (!atomic_read(&nmi_running))
atomic_read       164 arch/sh/kernel/ftrace.c 	} while (atomic_read(&nmi_running));
atomic_read       330 arch/sh/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read        50 arch/sh/kernel/irq.c 	seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count));
atomic_read       138 arch/sh/kernel/perf_event.c 		if (atomic_read(&num_events) == 0 &&
atomic_read       364 arch/sh/kernel/smp.c 	if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
atomic_read       396 arch/sh/kernel/smp.c 	if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
atomic_read       439 arch/sh/kernel/smp.c 	if ((atomic_read(&vma->vm_mm->mm_users) != 1) ||
atomic_read       130 arch/sparc/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       128 arch/sparc/kernel/kgdb_64.c 	if (atomic_read(&kgdb_active) != -1)
atomic_read        57 arch/sparc/kernel/nmi.c 	if (atomic_read(&nmi_active)) {
atomic_read       167 arch/sparc/kernel/nmi.c 	if (!atomic_read(&nmi_active))
atomic_read       193 arch/sparc/kernel/nmi.c 	if (!atomic_read(&nmi_active)) {
atomic_read       287 arch/sparc/kernel/nmi.c 	if (atomic_read(&nmi_active) == -1) {
atomic_read       310 arch/sparc/kernel/nmi.c 	if (atomic_read(&nmi_active) == -1)
atomic_read      1177 arch/sparc/kernel/perf_event.c 	if (atomic_read(&active_events) == 0) {
atomic_read      1178 arch/sparc/kernel/perf_event.c 		if (atomic_read(&nmi_active) > 0) {
atomic_read      1180 arch/sparc/kernel/perf_event.c 			BUG_ON(atomic_read(&nmi_active) != 0);
atomic_read      1190 arch/sparc/kernel/perf_event.c 		if (atomic_read(&nmi_active) == 0)
atomic_read      1425 arch/sparc/kernel/perf_event.c 	if (atomic_read(&nmi_active) < 0)
atomic_read      1624 arch/sparc/kernel/perf_event.c 	if (!atomic_read(&active_events))
atomic_read      1085 arch/sparc/kernel/smp_64.c 	if (atomic_read(&mm->mm_users) == 1) {
atomic_read      1123 arch/sparc/kernel/smp_64.c 	if (mm == current->mm && atomic_read(&mm->mm_users) == 1)
atomic_read      1139 arch/sparc/kernel/smp_64.c 	if (mm == current->mm && atomic_read(&mm->mm_users) == 1)
atomic_read      1184 arch/sparc/kernel/smp_64.c 		while (atomic_read(&smp_capture_registry) != ncpus)
atomic_read      2025 arch/sparc/kernel/traps_64.c 	if ((cnt = atomic_read(ocnt)) != 0) {
atomic_read       577 arch/sparc/mm/init_64.c 		   atomic_read(&dcpage_flushes));
atomic_read       580 arch/sparc/mm/init_64.c 		   atomic_read(&dcpage_flushes_xcall));
atomic_read      1693 arch/sparc/mm/srmmu.c 			if (atomic_read(&mm->mm_users) == 1 && current->active_mm == mm)
atomic_read        61 arch/sparc/oprofile/init.c 	if (atomic_read(&nmi_active) <= 0)
atomic_read       109 arch/um/drivers/port_kern.c 	if (atomic_read(&port->wait_count) == 0) {
atomic_read       323 arch/um/kernel/process.c 	return atomic_read(&using_sysemu);
atomic_read       528 arch/um/kernel/tlb.c 	if (atomic_read(&current->mm->mm_users) == 0)
atomic_read       556 arch/um/kernel/tlb.c 	if (atomic_read(&mm->mm_users) == 0)
atomic_read       342 arch/x86/events/core.c 		if (atomic_read(&pmc_refcount) == 0) {
atomic_read       383 arch/x86/events/core.c 			if (i != what && atomic_read(&x86_pmu.lbr_exclusive[i]))
atomic_read      1562 arch/x86/events/core.c 	if (!atomic_read(&active_events))
atomic_read      2665 arch/x86/events/intel/core.c 	if (!atomic_read(&era->ref) || era->config == reg->config) {
atomic_read       161 arch/x86/events/intel/uncore.c 	if (!atomic_read(&er->ref) ||
atomic_read       563 arch/x86/events/intel/uncore_nhmex.c 		if (!atomic_read(&er->ref) || er->config == config) {
atomic_read       590 arch/x86/events/intel/uncore_nhmex.c 	if (__BITS_VALUE(atomic_read(&er->ref), idx, 8)) {
atomic_read       597 arch/x86/events/intel/uncore_nhmex.c 	if (!atomic_read(&er->ref) || !((er->config ^ config) & mask)) {
atomic_read      1004 arch/x86/events/intel/uncore_nhmex.c 		if (!atomic_read(&er->ref) || er->config == reg1->config) {
atomic_read      1015 arch/x86/events/intel/uncore_nhmex.c 		if (!__BITS_VALUE(atomic_read(&er->ref), idx - 2, 8) ||
atomic_read      1023 arch/x86/events/intel/uncore_nhmex.c 		if (!atomic_read(&er->ref) ||
atomic_read       878 arch/x86/events/intel/uncore_snbep.c 		if (!__BITS_VALUE(atomic_read(&er->ref), i, 6) ||
atomic_read      1005 arch/x86/events/intel/uncore_snbep.c 	if (!__BITS_VALUE(atomic_read(&er->ref), idx, 8) ||
atomic_read        34 arch/x86/include/asm/mmu_context.h 	    atomic_read(&mm->context.perf_rdpmc_allowed))
atomic_read        25 arch/x86/include/asm/qspinlock.h 	val |= atomic_read(&lock->val) & ~_Q_PENDING_MASK;
atomic_read        98 arch/x86/include/asm/qspinlock.h 		while (atomic_read(&lock->val) != 0)
atomic_read       101 arch/x86/include/asm/refcount.h 	c = atomic_read(&(r->refs));
atomic_read       852 arch/x86/include/asm/uv/uv_bau.h 	if (atomic_read(v) >= u) {
atomic_read       406 arch/x86/kernel/apic/apic.c 	rsvd = atomic_read(&eilvt_offsets[offset]);
atomic_read       629 arch/x86/kernel/cpu/mce/core.c 	if (atomic_read(&num_notifiers) > NUM_DEFAULT_NOTIFIERS)
atomic_read       857 arch/x86/kernel/cpu/mce/core.c 	if (atomic_read(&mce_panicked))
atomic_read       978 arch/x86/kernel/cpu/mce/core.c 	while (atomic_read(&mce_callin) != cpus) {
atomic_read      1004 arch/x86/kernel/cpu/mce/core.c 		while (atomic_read(&mce_executing) < order) {
atomic_read      1017 arch/x86/kernel/cpu/mce/core.c 	*no_way_out = atomic_read(&global_nwo);
atomic_read      1049 arch/x86/kernel/cpu/mce/core.c 		while (atomic_read(&mce_executing) <= cpus) {
atomic_read      1063 arch/x86/kernel/cpu/mce/core.c 		while (atomic_read(&mce_executing) != 0) {
atomic_read       200 arch/x86/kernel/cpu/mce/intel.c 		if (!atomic_read(&cmci_storm_on_cpus)) {
atomic_read       289 arch/x86/kernel/cpu/mce/therm_throt.c 	if (!atomic_read(&therm_throt_en))
atomic_read       535 arch/x86/kernel/cpu/microcode/core.c 	while (atomic_read(t) < all_cpus) {
atomic_read       538 arch/x86/kernel/cpu/microcode/core.c 				all_cpus - atomic_read(t));
atomic_read      2222 arch/x86/kernel/cpu/resctrl/rdtgroup.c 		if (atomic_read(&sentry->waitcount) != 0)
atomic_read      2264 arch/x86/kernel/cpu/resctrl/rdtgroup.c 		if (atomic_read(&rdtgrp->waitcount) != 0)
atomic_read      1042 arch/x86/kernel/ftrace.c 	if (unlikely(atomic_read(&current->tracing_graph_pause)))
atomic_read       163 arch/x86/kernel/irq.c 	seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count));
atomic_read       165 arch/x86/kernel/irq.c 	seq_printf(p, "%*s: %10u\n", prec, "MIS", atomic_read(&irq_mis_count));
atomic_read       223 arch/x86/kernel/irq.c 	u64 sum = atomic_read(&irq_err_count);
atomic_read       502 arch/x86/kernel/kgdb.c 		if (atomic_read(&kgdb_active) != -1) {
atomic_read       533 arch/x86/kernel/kgdb.c 		if (atomic_read(&kgdb_cpu_doing_single_step) != -1) {
atomic_read       873 arch/x86/kernel/reboot.c 	while ((atomic_read(&waiting_for_crash_ipi) > 0) && msecs) {
atomic_read       121 arch/x86/kernel/smp.c 	if (raw_smp_processor_id() == atomic_read(&stopping_cpu))
atomic_read       316 arch/x86/kernel/tboot.c 	while (atomic_read((atomic_t *)&tboot->num_in_wfs) != num_aps &&
atomic_read       325 arch/x86/kernel/tboot.c 	return !(atomic_read((atomic_t *)&tboot->num_in_wfs) == num_aps);
atomic_read       332 arch/x86/kernel/tboot.c 		if (tboot_wait_for_aps(atomic_read(&ap_wfs_count)))
atomic_read       576 arch/x86/kernel/traps.c 	if (unlikely(atomic_read(&modifying_ftrace_code)) &&
atomic_read       334 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&start_count) != cpus - 1) {
atomic_read       335 arch/x86/kernel/tsc_sync.c 		if (atomic_read(&skip_test) > 0) {
atomic_read       349 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&stop_count) != cpus-1)
atomic_read       393 arch/x86/kernel/tsc_sync.c 	if (atomic_read(&test_runs) > 0)
atomic_read       431 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&start_count) != cpus)
atomic_read       449 arch/x86/kernel/tsc_sync.c 	while (atomic_read(&stop_count) != cpus)
atomic_read       462 arch/x86/kernel/tsc_sync.c 	if (!atomic_read(&test_runs))
atomic_read       182 arch/x86/kvm/hyperv.c 	gsi = atomic_read(&synic->sint_to_gsi[sint]);
atomic_read      1338 arch/x86/kvm/hyperv.c 	if (likely(!atomic_read(&hv->num_mismatched_vp_indexes))) {
atomic_read       247 arch/x86/kvm/i8254.c 	if (atomic_read(&ps->reinject) && !atomic_xchg(&ps->irq_ack, 0))
atomic_read       262 arch/x86/kvm/i8254.c 	if (atomic_read(&kvm->arch.vapics_in_nmi_mode) > 0)
atomic_read       272 arch/x86/kvm/i8254.c 	if (atomic_read(&ps->reinject))
atomic_read       295 arch/x86/kvm/i8254.c 	if (atomic_read(&ps->reinject) == reinject)
atomic_read      1552 arch/x86/kvm/lapic.c 	if (atomic_read(&apic->lapic_timer.pending))
atomic_read      1736 arch/x86/kvm/lapic.c 		if (atomic_read(&ktimer->pending)) {
atomic_read      1756 arch/x86/kvm/lapic.c 	if (!apic_lvtt_period(apic) && atomic_read(&ktimer->pending))
atomic_read      1770 arch/x86/kvm/lapic.c 	if (!apic_lvtt_period(apic) && atomic_read(&apic->lapic_timer.pending))
atomic_read      2233 arch/x86/kvm/lapic.c 		return atomic_read(&apic->lapic_timer.pending);
atomic_read      2357 arch/x86/kvm/lapic.c 	if (atomic_read(&apic->lapic_timer.pending) > 0) {
atomic_read      5369 arch/x86/kvm/mmu.c 	return atomic_read(&sp->write_flooding_count) >= 3;
atomic_read      1851 arch/x86/kvm/x86.c 			 atomic_read(&vcpu->kvm->online_vcpus));
atomic_read      1866 arch/x86/kvm/x86.c 			    atomic_read(&vcpu->kvm->online_vcpus),
atomic_read      2229 arch/x86/kvm/x86.c 			atomic_read(&kvm->online_vcpus));
atomic_read      7230 arch/x86/kvm/x86.c 	    atomic_read(&kvm_guest_has_master_clock) != 0)
atomic_read      9157 arch/x86/kvm/x86.c 	if (kvm_check_tsc_unstable() && atomic_read(&kvm->online_vcpus) != 0)
atomic_read      9608 arch/x86/kvm/x86.c 	for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
atomic_read      10278 arch/x86/kvm/x86.c 	return atomic_read(&kvm->arch.assigned_device_count);
atomic_read      10296 arch/x86/kvm/x86.c 	return atomic_read(&kvm->arch.noncoherent_dma_count);
atomic_read        78 arch/x86/mm/mmio-mod.c 	return atomic_read(&mmiotrace_enabled);
atomic_read       479 arch/x86/platform/uv/uv_nmi.c 		nmi = atomic_read(&hub_nmi->in_nmi);
atomic_read       507 arch/x86/platform/uv/uv_nmi.c 			nmi = atomic_read(&hub_nmi->in_nmi);
atomic_read       517 arch/x86/platform/uv/uv_nmi.c 			nmi = atomic_read(&uv_in_nmi);
atomic_read       539 arch/x86/platform/uv/uv_nmi.c 	if (cpu == atomic_read(&hub_nmi->cpu_owner)) {
atomic_read       657 arch/x86/platform/uv/uv_nmi.c 		atomic_read(&uv_nmi_cpus_in_nmi), num_online_cpus());
atomic_read       725 arch/x86/platform/uv/uv_nmi.c 		while (atomic_read(&uv_nmi_cpus_in_nmi) > 0)
atomic_read       729 arch/x86/platform/uv/uv_nmi.c 		while (atomic_read(&uv_nmi_slave_continue))
atomic_read       738 arch/x86/platform/uv/uv_nmi.c 		int in = atomic_read(&uv_nmi_cpus_in_nmi);
atomic_read       744 arch/x86/platform/uv/uv_nmi.c 		while (!atomic_read(&uv_nmi_slave_continue))
atomic_read       760 arch/x86/platform/uv/uv_nmi.c 			atomic_read(&uv_nmi_cpus_in_nmi), cpu);
atomic_read       778 arch/x86/platform/uv/uv_nmi.c 		while (!atomic_read(&uv_nmi_slave_continue))
atomic_read       815 arch/x86/platform/uv/uv_nmi.c 	while (atomic_read(&uv_nmi_kexec_failed) == 0)
atomic_read       875 arch/x86/platform/uv/uv_nmi.c 			sig = atomic_read(&uv_nmi_slave_continue);
atomic_read       911 arch/x86/platform/uv/uv_nmi.c 	master = (atomic_read(&uv_nmi_cpu) == cpu);
atomic_read        50 arch/x86/xen/spinlock.c 	if (atomic_read(nest_cnt) == 1 && xen_test_irq_pending(irq)) {
atomic_read        47 arch/xtensa/kernel/jump_label.c 		while (atomic_read(&patch->cpu_count) <= num_online_cpus())
atomic_read      5240 block/bfq-iosched.c 	if (atomic_read(&bic->icq.ioc->active_ref) == 0 ||
atomic_read       603 block/bio.c    		BIO_BUG_ON(!atomic_read(&bio->__bi_cnt));
atomic_read      1814 block/bio.c    	BUG_ON(atomic_read(&bio->__bi_remaining) <= 0);
atomic_read       974 block/blk-cgroup.c 		if (blkcg_debug_stats && atomic_read(&blkg->use_delay)) {
atomic_read       978 block/blk-cgroup.c 					 atomic_read(&blkg->use_delay),
atomic_read      1291 block/blk-cgroup.c 		if (ioc && atomic_read(&ioc->nr_tasks) > 1)
atomic_read      1660 block/blk-cgroup.c 		int cur_use = atomic_read(&blkg->use_delay);
atomic_read      1701 block/blk-cgroup.c 		if (atomic_read(&blkg->use_delay)) {
atomic_read       965 block/blk-iocost.c 	ioc_gen = atomic_read(&ioc->hweight_gen);
atomic_read      1092 block/blk-iocost.c 	iocg->hweight_gen = atomic_read(&ioc->hweight_gen) - 1;
atomic_read      1237 block/blk-iocost.c 	if (!atomic_read(&blkg->use_delay) &&
atomic_read      1421 block/blk-iocost.c 		    !atomic_read(&iocg_to_blkg(iocg)->use_delay) &&
atomic_read        99 block/blk-iolatency.c 	return atomic_read(&blkiolat->enabled) > 0;
atomic_read       286 block/blk-iolatency.c 	unsigned use_delay = atomic_read(&lat_to_blkg(iolat)->use_delay);
atomic_read       329 block/blk-iolatency.c 	unsigned long old = atomic_read(&lat_info->scale_cookie);
atomic_read       396 block/blk-iolatency.c 	unsigned int our_cookie = atomic_read(&iolat->scale_cookie);
atomic_read       409 block/blk-iolatency.c 	cur_cookie = atomic_read(&lat_info->scale_cookie);
atomic_read       549 block/blk-iolatency.c 	    atomic_read(&lat_info->scale_cookie) == DEFAULT_SCALE_COOKIE)
atomic_read       683 block/blk-iolatency.c 		cookie = atomic_read(&lat_info->scale_cookie);
atomic_read       992 block/blk-iolatency.c 			   atomic_read(&parent->child_lat.scale_cookie));
atomic_read        99 block/blk-mq-debugfs.c 	seq_printf(m, "%d\n", atomic_read(&q->pm_only));
atomic_read       449 block/blk-mq-debugfs.c 		   atomic_read(&tags->active_queues));
atomic_read       613 block/blk-mq-debugfs.c 	seq_printf(m, "%d\n", atomic_read(&hctx->nr_active));
atomic_read        87 block/blk-mq-tag.c 	users = atomic_read(&hctx->tags->active_queues);
atomic_read        95 block/blk-mq-tag.c 	return atomic_read(&hctx->nr_active) < depth;
atomic_read        11 block/blk-rq-qos.c 	unsigned int cur = atomic_read(v);
atomic_read       753 block/blk-wbt.c 			   atomic_read(&rwb->rq_wait[i].inflight));
atomic_read        82 block/blk-wbt.h 		ret += atomic_read(&rwb->rq_wait[i].inflight);
atomic_read       686 drivers/acpi/apei/ghes.c 		count = atomic_read(&cache->count);
atomic_read      1810 drivers/android/binder.c 	if (thread->is_dead && !atomic_read(&thread->tmp_ref)) {
atomic_read      5639 drivers/android/binder.c 			atomic_read(&thread->tmp_ref));
atomic_read      5835 drivers/android/binder.c 		int temp = atomic_read(&stats->bc[i]);
atomic_read      5845 drivers/android/binder.c 		int temp = atomic_read(&stats->br[i]);
atomic_read      5857 drivers/android/binder.c 		int created = atomic_read(&stats->obj_created[i]);
atomic_read      5858 drivers/android/binder.c 		int deleted = atomic_read(&stats->obj_deleted[i]);
atomic_read      6048 drivers/android/binder.c 	unsigned int log_cur = atomic_read(&log->cur);
atomic_read       713 drivers/atm/idt77252.c 	entries = atomic_read(&scq->used);
atomic_read       768 drivers/atm/idt77252.c 	TXPRINTK("%d entries in SCQ used (push).\n", atomic_read(&scq->used));
atomic_read       771 drivers/atm/idt77252.c 		card->name, atomic_read(&scq->used),
atomic_read       798 drivers/atm/idt77252.c 		 card->name, atomic_read(&scq->used), scq->next);
atomic_read      2545 drivers/atm/idt77252.c 		while (atomic_read(&vc->scq->used) > 0) {
atomic_read      2549 drivers/atm/idt77252.c 					card->name, atomic_read(&vc->scq->used));
atomic_read      2835 drivers/atm/iphase.c              printk("section_bip: %d\n", atomic_read(&stats->section_bip));
atomic_read      2836 drivers/atm/iphase.c              printk("line_bip   : %d\n", atomic_read(&stats->line_bip));
atomic_read      2837 drivers/atm/iphase.c              printk("path_bip   : %d\n", atomic_read(&stats->path_bip));
atomic_read      2838 drivers/atm/iphase.c              printk("line_febe  : %d\n", atomic_read(&stats->line_febe));
atomic_read      2839 drivers/atm/iphase.c              printk("path_febe  : %d\n", atomic_read(&stats->path_febe));
atomic_read      2840 drivers/atm/iphase.c              printk("corr_hcs   : %d\n", atomic_read(&stats->corr_hcs));
atomic_read      2841 drivers/atm/iphase.c              printk("uncorr_hcs : %d\n", atomic_read(&stats->uncorr_hcs));
atomic_read      2842 drivers/atm/iphase.c              printk("tx_cells   : %d\n", atomic_read(&stats->tx_cells));
atomic_read      2843 drivers/atm/iphase.c              printk("rx_cells   : %d\n", atomic_read(&stats->rx_cells));
atomic_read      3060 drivers/atm/iphase.c         if (atomic_read(&vcc->stats->tx) % 20 == 0) {
atomic_read        54 drivers/atm/suni.c     if (atomic_read(&stats->s) < 0) atomic_set(&stats->s,INT_MAX);
atomic_read       166 drivers/atm/uPD98402.c     if (atomic_read(&PRIV(dev)->sonet_stats.s) < 0) \
atomic_read      1142 drivers/auxdisplay/panel.c 	if (!atomic_read(&keypad_available)) {
atomic_read       488 drivers/base/dd.c 	if (local_trigger_count != atomic_read(&deferred_trigger_count))
atomic_read       495 drivers/base/dd.c 	int local_trigger_count = atomic_read(&deferred_trigger_count);
atomic_read       672 drivers/base/dd.c 		 atomic_read(&probe_count));
atomic_read       673 drivers/base/dd.c 	if (atomic_read(&probe_count))
atomic_read       688 drivers/base/dd.c 	wait_event(probe_waitqueue, atomic_read(&probe_count) == 0);
atomic_read       200 drivers/base/power/domain.c 	if (!WARN_ON(atomic_read(&genpd->sd_count) == 0))
atomic_read       516 drivers/base/power/domain.c 			atomic_read(&genpd->sd_count) > 0)
atomic_read       550 drivers/base/power/domain.c 		if (atomic_read(&genpd->sd_count) > 0)
atomic_read       967 drivers/base/power/domain.c 	    || atomic_read(&genpd->sd_count) > 0)
atomic_read      1352 drivers/base/power/main.c 				atomic_read(&dev->power.usage_count) > 1 ||
atomic_read       266 drivers/base/power/runtime.c 	else if (atomic_read(&dev->power.usage_count) > 0)
atomic_read       269 drivers/base/power/runtime.c 			atomic_read(&dev->power.child_count))
atomic_read      1391 drivers/base/power/runtime.c 	     atomic_read(&dev->power.child_count) > 0,
atomic_read      1718 drivers/base/power/runtime.c 	return atomic_read(&dev->power.usage_count) <= 1 &&
atomic_read      1719 drivers/base/power/runtime.c 		(atomic_read(&dev->power.child_count) == 0 ||
atomic_read       510 drivers/base/power/sysfs.c 	return sprintf(buf, "%d\n", atomic_read(&dev->power.usage_count));
atomic_read       519 drivers/base/power/sysfs.c 		0 : atomic_read(&dev->power.child_count));
atomic_read        51 drivers/base/power/wakeup.c 	unsigned int comb = atomic_read(&combined_event_count);
atomic_read       872 drivers/base/power/wakeup.c 	return ret || atomic_read(&pm_abort_suspend) > 0;
atomic_read        32 drivers/base/test/test_async_driver_probe.c 	if (atomic_read(&timeout)) {
atomic_read       247 drivers/base/test/test_async_driver_probe.c 	if (atomic_read(&async_completed) != async_id) {
atomic_read       251 drivers/base/test/test_async_driver_probe.c 	} else if (!atomic_read(&errors) && !atomic_read(&warnings)) {
atomic_read       279 drivers/base/test/test_async_driver_probe.c 	       atomic_read(&errors), atomic_read(&warnings));
atomic_read       181 drivers/block/aoe/aoecmd.c 	if (skb && atomic_read(&skb_shinfo(skb)->dataref) == 1) {
atomic_read       235 drivers/block/aoe/aoecmd.c 	if (atomic_read(&skb_shinfo(skb)->dataref) != 1) {
atomic_read       412 drivers/block/aoe/aoedev.c 	while (atomic_read(&skb_shinfo(skb)->dataref) != 1 && i-- > 0)
atomic_read       181 drivers/block/drbd/drbd_actlog.c 	D_ASSERT(device, atomic_read(&device->md_io.in_use) == 1);
atomic_read       249 drivers/block/drbd/drbd_actlog.c 	D_ASSERT(device, atomic_read(&device->local_cnt) > 0);
atomic_read       268 drivers/block/drbd/drbd_actlog.c 	D_ASSERT(device, atomic_read(&device->local_cnt) > 0);
atomic_read       657 drivers/block/drbd/drbd_actlog.c 	D_ASSERT(device, atomic_read(&device->local_cnt));
atomic_read      1164 drivers/block/drbd/drbd_bitmap.c 	if (atomic_read(&ctx->in_flight))
atomic_read       142 drivers/block/drbd/drbd_debugfs.c 		if (atomic_read(&tmp.in_use)) {
atomic_read       166 drivers/block/drbd/drbd_debugfs.c 		int n = atomic_read(&device->ap_actlog_cnt);
atomic_read       203 drivers/block/drbd/drbd_debugfs.c 		in_flight = atomic_read(&ctx->in_flight);
atomic_read      2029 drivers/block/drbd/drbd_int.h 	if (atomic_read(&device->which) < 0)				\
atomic_read      2032 drivers/block/drbd/drbd_int.h 			atomic_read(&device->which))
atomic_read      2262 drivers/block/drbd/drbd_int.h 	if (atomic_read(&device->suspend_cnt))
atomic_read      2275 drivers/block/drbd/drbd_int.h 	if (atomic_read(&device->ap_bio_cnt) > mxb)
atomic_read      2299 drivers/block/drbd/drbd_main.c 			expect(atomic_read(&req->completion_ref) == 0) &&
atomic_read      2306 drivers/block/drbd/drbd_main.c 				req, atomic_read(&req->completion_ref),
atomic_read      2746 drivers/block/drbd/drbd_main.c 	if (atomic_read(&connection->current_epoch->epoch_size) !=  0)
atomic_read      2747 drivers/block/drbd/drbd_main.c 		drbd_err(connection, "epoch_size:%d\n", atomic_read(&connection->current_epoch->epoch_size));
atomic_read      3574 drivers/block/drbd/drbd_main.c 		int cnt = atomic_read(&device->ap_bio_cnt);
atomic_read      3638 drivers/block/drbd/drbd_main.c 	if (flags == BM_LOCKED_CHANGE_ALLOWED || atomic_read(&device->ap_bio_cnt) == 0) {
atomic_read       715 drivers/block/drbd/drbd_nl.c 	wait_event(device->misc_wait, atomic_read(&device->ap_pending_cnt) == 0);
atomic_read       911 drivers/block/drbd/drbd_nl.c 	wait_event(device->misc_wait, !atomic_read(&device->ap_bio_cnt));
atomic_read      1559 drivers/block/drbd/drbd_nl.c 	if (atomic_read(&device->ap_bio_cnt))
atomic_read      1969 drivers/block/drbd/drbd_nl.c 	wait_event(device->misc_wait, !atomic_read(&device->ap_pending_cnt) || drbd_suspended(device));
atomic_read      3399 drivers/block/drbd/drbd_nl.c 	s->dev_upper_pending = atomic_read(&device->ap_bio_cnt);
atomic_read      3400 drivers/block/drbd/drbd_nl.c 	s->dev_lower_pending = atomic_read(&device->local_cnt);
atomic_read      3644 drivers/block/drbd/drbd_nl.c 	s->peer_dev_pending = atomic_read(&device->ap_pending_cnt) +
atomic_read      3645 drivers/block/drbd/drbd_nl.c 			      atomic_read(&device->rs_pending_cnt);
atomic_read      3646 drivers/block/drbd/drbd_nl.c 	s->peer_dev_unacked = atomic_read(&device->unacked_cnt);
atomic_read      3842 drivers/block/drbd/drbd_nl.c 	    nla_put_u32(skb, T_ap_bio_cnt, atomic_read(&device->ap_bio_cnt)) ||
atomic_read      3843 drivers/block/drbd/drbd_nl.c 	    nla_put_u32(skb, T_ap_pending_cnt, atomic_read(&device->ap_pending_cnt)) ||
atomic_read      3844 drivers/block/drbd/drbd_nl.c 	    nla_put_u32(skb, T_rs_pending_cnt, atomic_read(&device->rs_pending_cnt)))
atomic_read       294 drivers/block/drbd/drbd_proc.c 			   atomic_read(&device->local_cnt),
atomic_read       295 drivers/block/drbd/drbd_proc.c 			   atomic_read(&device->ap_pending_cnt) +
atomic_read       296 drivers/block/drbd/drbd_proc.c 			   atomic_read(&device->rs_pending_cnt),
atomic_read       297 drivers/block/drbd/drbd_proc.c 			   atomic_read(&device->unacked_cnt),
atomic_read       298 drivers/block/drbd/drbd_proc.c 			   atomic_read(&device->ap_bio_cnt),
atomic_read       319 drivers/block/drbd/drbd_proc.c 			seq_printf(seq, "\tblocked on activity log: %d\n", atomic_read(&device->ap_actlog_cnt));
atomic_read       228 drivers/block/drbd/drbd_receiver.c 		if (!atomic_read(&device->pp_in_use_by_net))
atomic_read       274 drivers/block/drbd/drbd_receiver.c 	if (atomic_read(&device->pp_in_use) < mxb)
atomic_read       279 drivers/block/drbd/drbd_receiver.c 	if (page && atomic_read(&device->pp_in_use_by_net) > 512)
atomic_read       287 drivers/block/drbd/drbd_receiver.c 		if (atomic_read(&device->pp_in_use) < mxb) {
atomic_read       411 drivers/block/drbd/drbd_receiver.c 	D_ASSERT(device, atomic_read(&peer_req->pending_bios) == 0);
atomic_read      1368 drivers/block/drbd/drbd_receiver.c 		epoch_size = atomic_read(&epoch->epoch_size);
atomic_read      1383 drivers/block/drbd/drbd_receiver.c 		    atomic_read(&epoch->active) == 0 &&
atomic_read      1806 drivers/block/drbd/drbd_receiver.c 		if (atomic_read(&connection->current_epoch->epoch_size)) {
atomic_read      1824 drivers/block/drbd/drbd_receiver.c 	if (atomic_read(&connection->current_epoch->epoch_size)) {
atomic_read      2805 drivers/block/drbd/drbd_receiver.c 			atomic_read(&device->rs_sect_ev);
atomic_read      2807 drivers/block/drbd/drbd_receiver.c 	if (atomic_read(&device->ap_actlog_cnt)
atomic_read      5281 drivers/block/drbd/drbd_receiver.c 	i = atomic_read(&device->pp_in_use_by_net);
atomic_read      5284 drivers/block/drbd/drbd_receiver.c 	i = atomic_read(&device->pp_in_use);
atomic_read      5890 drivers/block/drbd/drbd_receiver.c 		    atomic_read(&device->ap_in_flight) == 0 &&
atomic_read        98 drivers/block/drbd/drbd_req.c 		atomic_read(&req->completion_ref) ||
atomic_read       102 drivers/block/drbd/drbd_req.c 				s, atomic_read(&req->completion_ref));
atomic_read       262 drivers/block/drbd/drbd_req.c 	    req->epoch == atomic_read(&first_peer_device(device)->connection->current_tle_nr))
atomic_read       930 drivers/block/drbd/drbd_req.c 		return atomic_read(&device->local_cnt) >
atomic_read       931 drivers/block/drbd/drbd_req.c 			atomic_read(&device->ap_pending_cnt) + atomic_read(&device->rs_pending_cnt);
atomic_read      1016 drivers/block/drbd/drbd_req.c 	    atomic_read(&device->ap_in_flight) >= nc->cong_fill) {
atomic_read      1364 drivers/block/drbd/drbd_req.c 	req->epoch = atomic_read(&first_peer_device(device)->connection->current_tle_nr);
atomic_read      1701 drivers/block/drbd/drbd_worker.c 	if (atomic_read(&device->unacked_cnt) || atomic_read(&device->rs_pending_cnt)) {
atomic_read      2125 drivers/block/drbd/drbd_worker.c 			atomic_read(&connection->current_tle_nr) !=
atomic_read       880 drivers/block/floppy.c 	if (WARN(atomic_read(&usage_count) == 0,
atomic_read      2899 drivers/block/floppy.c 	if (WARN(atomic_read(&usage_count) == 0,
atomic_read      4202 drivers/block/floppy.c 		if (WARN(atomic_read(&usage_count) == 0,
atomic_read      4726 drivers/block/floppy.c 	if (atomic_read(&usage_count))
atomic_read      4986 drivers/block/floppy.c 	if (atomic_read(&usage_count))
atomic_read      1246 drivers/block/loop.c 	if (atomic_read(&lo->lo_refcnt) > 1) {
atomic_read      2202 drivers/block/loop.c 		if (atomic_read(&lo->lo_refcnt) > 0) {
atomic_read       762 drivers/block/mtip32xx/mtip32xx.c 			WARN_ON_ONCE(atomic_read(&dd->irq_workers_active) != 0);
atomic_read      2675 drivers/block/mtip32xx/mtip32xx.c 			} while (atomic_read(&dd->irq_workers_active) != 0 &&
atomic_read      2678 drivers/block/mtip32xx/mtip32xx.c 			if (atomic_read(&dd->irq_workers_active) != 0)
atomic_read      4205 drivers/block/mtip32xx/mtip32xx.c 	} while (atomic_read(&dd->irq_workers_active) != 0 &&
atomic_read      4211 drivers/block/mtip32xx/mtip32xx.c 	if (atomic_read(&dd->irq_workers_active) != 0) {
atomic_read       401 drivers/block/nbd.c 				    atomic_read(&config->live_connections),
atomic_read       867 drivers/block/nbd.c 				  atomic_read(&config->live_connections) > 0,
atomic_read      1308 drivers/block/nbd.c 					 atomic_read(&config->recv_threads) == 0);
atomic_read       747 drivers/block/paride/pt.c 	if (atomic_read(&tape->available) > 1)
atomic_read       511 drivers/block/pktcdvd.c 	BUG_ON(atomic_read(&pd->cdrw.pending_bios) <= 0);
atomic_read       844 drivers/block/pktcdvd.c 	if (atomic_read(&pd->iosched.attention) == 0)
atomic_read       869 drivers/block/pktcdvd.c 				if (atomic_read(&pd->cdrw.pending_bios) > 0) {
atomic_read       878 drivers/block/pktcdvd.c 				if (atomic_read(&pd->cdrw.pending_bios) > 0) {
atomic_read      1338 drivers/block/pktcdvd.c 			if (atomic_read(&pkt->io_wait) > 0)
atomic_read      1341 drivers/block/pktcdvd.c 			if (atomic_read(&pkt->io_errors) > 0) {
atomic_read      1349 drivers/block/pktcdvd.c 			if (atomic_read(&pkt->io_wait) > 0)
atomic_read      1387 drivers/block/pktcdvd.c 		if (atomic_read(&pkt->run_sm) > 0) {
atomic_read      1447 drivers/block/pktcdvd.c 			if (atomic_read(&pd->scan_queue) > 0)
atomic_read      1452 drivers/block/pktcdvd.c 				if (atomic_read(&pkt->run_sm) > 0)
atomic_read      1457 drivers/block/pktcdvd.c 			if (atomic_read(&pd->iosched.attention) != 0)
atomic_read      2551 drivers/block/pktcdvd.c 	seq_printf(m, "\tbios pending:\t\t%d\n", atomic_read(&pd->cdrw.pending_bios));
atomic_read       147 drivers/block/rsxx/core.c 			i, atomic_read(&card->ctrl[i].stats.hw_q_depth));
atomic_read       126 drivers/block/rsxx/dev.c 		if (atomic_read(&meta->error))
atomic_read       201 drivers/block/rsxx/dma.c 		q_depth += atomic_read(&card->ctrl[i].stats.hw_q_depth);
atomic_read       348 drivers/block/rsxx/dma.c 	if (atomic_read(&ctrl->stats.hw_q_depth) == 0 ||
atomic_read       565 drivers/block/rsxx/dma.c 	if (atomic_read(&ctrl->stats.hw_q_depth) == 0)
atomic_read       746 drivers/block/xen-blkback/blkback.c 	if (atomic_dec_and_test(&ring->inflight) && atomic_read(&blkif->drain)) {
atomic_read      1067 drivers/block/xen-blkback/blkback.c 		if (atomic_read(&ring->inflight) == 0)
atomic_read      1072 drivers/block/xen-blkback/blkback.c 		if (!atomic_read(&blkif->drain))
atomic_read       267 drivers/block/xen-blkback/xenbus.c 		if (atomic_read(&ring->inflight) > 0) {
atomic_read       299 drivers/block/xen-blkback/xenbus.c 		BUG_ON(atomic_read(&ring->persistent_gnt_in_use) != 0);
atomic_read       146 drivers/bluetooth/bcm203x.c 	if (atomic_read(&data->shutdown))
atomic_read       160 drivers/bluetooth/bfusb.c 		while ((atomic_read(&data->pending_tx) < BFUSB_MAX_BULK_TX) &&
atomic_read        76 drivers/char/agp/backend.c 	if (atomic_read(&bridge->agp_in_use))
atomic_read       752 drivers/char/agp/frontend.c         if (atomic_read(&agp_bridge->agp_in_use))
atomic_read       229 drivers/char/agp/generic.c 	cur_memory = atomic_read(&bridge->current_memory_agp);
atomic_read       382 drivers/char/agp/generic.c 	info->current_memory = atomic_read(&bridge->current_memory_agp);
atomic_read       518 drivers/char/apm-emulation.c 	if (atomic_read(&userspace_notification_inhibit))
atomic_read       556 drivers/char/apm-emulation.c 			atomic_read(&suspend_acks_pending) == 0,
atomic_read       165 drivers/char/ipmi/ipmb_dev_int.c 	if (atomic_read(&ipmb_dev->request_queue_len))
atomic_read       184 drivers/char/ipmi/ipmb_dev_int.c 	if (atomic_read(&ipmb_dev->request_queue_len) >=
atomic_read       651 drivers/char/ipmi/ipmi_msghandler.c 	((unsigned int) atomic_read(&(intf)->stats[IPMI_STAT_ ## stat]))
atomic_read      4764 drivers/char/ipmi/ipmi_msghandler.c 	if (atomic_read(&stop_operation))
atomic_read      4769 drivers/char/ipmi/ipmi_msghandler.c 		if (atomic_read(&intf->event_waiters)) {
atomic_read      4885 drivers/char/ipmi/ipmi_msghandler.c 	while (atomic_read(&panic_done_count) != 0)
atomic_read      5203 drivers/char/ipmi/ipmi_msghandler.c 		count = atomic_read(&smi_msg_inuse_count);
atomic_read      5206 drivers/char/ipmi/ipmi_msghandler.c 		count = atomic_read(&recv_msg_inuse_count);
atomic_read       159 drivers/char/ipmi/ipmi_poweroff.c 	while (atomic_read(&dummy_count) > 0) {
atomic_read       250 drivers/char/ipmi/ipmi_si_intf.c 	((unsigned int) atomic_read(&(smi)->stats[SI_STAT_ ## stat]))
atomic_read       833 drivers/char/ipmi/ipmi_si_intf.c 	    && (atomic_read(&smi_info->req_events))) {
atomic_read      1028 drivers/char/ipmi/ipmi_si_intf.c 			if (atomic_read(&smi_info->need_watch)) {
atomic_read       302 drivers/char/ipmi/ipmi_ssif.c 	((unsigned int) atomic_read(&(ssif)->stats[SSIF_STAT_ ## stat]))
atomic_read       525 drivers/char/ipmi/ipmi_watchdog.c 	while (atomic_read(&panic_done_count) != 0)
atomic_read       538 drivers/char/ipmi/ipmi_watchdog.c 	while (atomic_read(&panic_done_count) != 0)
atomic_read      1080 drivers/char/ipmi/ipmi_watchdog.c 	while (atomic_read(&msg_tofree))
atomic_read       612 drivers/char/ppdev.c 		ret = atomic_read(&pp->irqc);
atomic_read       782 drivers/char/ppdev.c 	if (atomic_read(&pp->irqc))
atomic_read       615 drivers/clocksource/arm_arch_timer.c 	return atomic_read(&timer_unstable_counter_workaround_in_use);
atomic_read        69 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read        98 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       123 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       160 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       184 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       216 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       242 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       275 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       317 drivers/connector/cn_proc.c 	if (atomic_read(&proc_event_num_listeners) < 1)
atomic_read       139 drivers/connector/cn_queue.c 	while (atomic_read(&dev->refcnt)) {
atomic_read       141 drivers/connector/cn_queue.c 		       dev->name, atomic_read(&dev->refcnt));
atomic_read       155 drivers/cpuidle/coupled.c 	while (atomic_read(a) < n)
atomic_read       163 drivers/cpuidle/coupled.c 	while (atomic_read(a) > n)
atomic_read       245 drivers/cpuidle/coupled.c 	int r = atomic_read(&coupled->ready_waiting_counts) >> WAITING_BITS;
atomic_read       257 drivers/cpuidle/coupled.c 	int r = atomic_read(&coupled->ready_waiting_counts) >> WAITING_BITS;
atomic_read       269 drivers/cpuidle/coupled.c 	int w = atomic_read(&coupled->ready_waiting_counts) & WAITING_MASK;
atomic_read       281 drivers/cpuidle/coupled.c 	int w = atomic_read(&coupled->ready_waiting_counts) & WAITING_MASK;
atomic_read       234 drivers/crypto/atmel-ecc.c 		tfm_cnt = atomic_read(&i2c_priv->tfm_count);
atomic_read       363 drivers/crypto/atmel-ecc.c 	if (atomic_read(&i2c_priv->tfm_count)) {
atomic_read       123 drivers/crypto/atmel-sha204a.c 	if (atomic_read(&i2c_priv->tfm_count)) {
atomic_read       374 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->session_count));
atomic_read       377 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->stream_count));
atomic_read       380 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->setkey_cnt[SPU_OP_CIPHER]));
atomic_read       383 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->op_counts[SPU_OP_CIPHER]));
atomic_read       386 drivers/crypto/bcm/util.c 			op_cnt = atomic_read(&ipriv->cipher_cnt[alg][mode]);
atomic_read       397 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->op_counts[SPU_OP_HASH]));
atomic_read       399 drivers/crypto/bcm/util.c 		op_cnt = atomic_read(&ipriv->hash_cnt[alg]);
atomic_read       409 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->setkey_cnt[SPU_OP_HMAC]));
atomic_read       412 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->op_counts[SPU_OP_HMAC]));
atomic_read       414 drivers/crypto/bcm/util.c 		op_cnt = atomic_read(&ipriv->hmac_cnt[alg]);
atomic_read       424 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->setkey_cnt[SPU_OP_AEAD]));
atomic_read       428 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->op_counts[SPU_OP_AEAD]));
atomic_read       430 drivers/crypto/bcm/util.c 		op_cnt = atomic_read(&ipriv->aead_cnt[alg]);
atomic_read       446 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->mb_no_spc));
atomic_read       449 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->mb_send_fail));
atomic_read       452 drivers/crypto/bcm/util.c 			       atomic_read(&ipriv->bad_icv));
atomic_read       151 drivers/crypto/caam/caamrng.c 	if (atomic_read(&bd->empty)) {
atomic_read       153 drivers/crypto/caam/caamrng.c 		if (atomic_read(&bd->empty) == BUF_EMPTY) {
atomic_read       164 drivers/crypto/caam/caamrng.c 		if (atomic_read(&bd->empty))
atomic_read       254 drivers/crypto/caam/caamrng.c 		if (atomic_read(&bd->empty) == BUF_PENDING)
atomic_read       132 drivers/crypto/caam/jr.c 	if (atomic_read(&jrpriv->tfm_count)) {
atomic_read       294 drivers/crypto/caam/jr.c 		tfm_cnt = atomic_read(&jrpriv->tfm_count);
atomic_read       279 drivers/crypto/cavium/nitrox/nitrox_dev.h 	return atomic_read(&ndev->state) == __NDEV_READY;
atomic_read       284 drivers/crypto/cavium/nitrox/nitrox_dev.h 	return atomic_read(&vfdev->state) == __NDEV_READY;
atomic_read       321 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 	if (!atomic_read(&cmdq->backlog_count))
atomic_read       542 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 	budget = atomic_read(&cmdq->pending_count);
atomic_read       549 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 		if (atomic_read(&sr->status) != REQ_POSTED)
atomic_read       601 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c 	if (atomic_read(&cmdq->backlog_count))
atomic_read        59 drivers/crypto/chelsio/chcr_core.c 	if (atomic_read(&dev->inflight)) {
atomic_read        63 drivers/crypto/chelsio/chcr_core.c 				atomic_read(&dev->inflight));
atomic_read        68 drivers/crypto/chelsio/chcr_core.c 				atomic_read(&dev->inflight));
atomic_read       249 drivers/crypto/chelsio/chcr_core.c 	if (atomic_read(&dev->inflight) != 0) {
atomic_read      1091 drivers/crypto/hisilicon/qm.c 	if (unlikely(atomic_read(&qp->qp_status.used) == QM_Q_DEPTH))
atomic_read      1332 drivers/crypto/hisilicon/qm.c 	while (atomic_read(&qp->qp_status.used)) {
atomic_read       846 drivers/crypto/hisilicon/sec/sec_drv.c 	return !atomic_read(&msg_ring->used);
atomic_read       866 drivers/crypto/hisilicon/sec/sec_drv.c 	if (write == read && atomic_read(&msg_ring->used) == SEC_QUEUE_LEN) {
atomic_read       888 drivers/crypto/hisilicon/sec/sec_drv.c 	return SEC_QUEUE_LEN - atomic_read(&msg_ring->used) >= num;
atomic_read       888 drivers/crypto/ixp4xx_crypto.c 	if (atomic_read(&ctx->configuring))
atomic_read       996 drivers/crypto/ixp4xx_crypto.c 	if (atomic_read(&ctx->configuring))
atomic_read       736 drivers/crypto/marvell/cesa.h 		u32 load = atomic_read(&engine->load);
atomic_read       425 drivers/crypto/qat/qat_common/adf_dev_mgr.c 	return atomic_read(&accel_dev->ref_count) != 0;
atomic_read        74 drivers/crypto/qat/qat_common/qat_crypto.c 		for (i = 0; i < atomic_read(&inst->refctr); i++)
atomic_read       108 drivers/crypto/qat/qat_common/qat_crypto.c 			ctr = atomic_read(&tmp_dev->ref_count);
atomic_read       135 drivers/crypto/qat/qat_common/qat_crypto.c 		ctr = atomic_read(&tmp_inst->refctr);
atomic_read       188 drivers/crypto/ux500/cryp/cryp_core.c 	ctx->session_id = atomic_read(&session_id);
atomic_read       407 drivers/crypto/ux500/cryp/cryp_core.c 		   ctx->session_id != atomic_read(&session_id)) {
atomic_read       117 drivers/crypto/virtio/virtio_crypto_mgr.c 	return atomic_read(&vcrypto_dev->ref_count) != 0;
atomic_read       198 drivers/crypto/virtio/virtio_crypto_mgr.c 			ctr = atomic_read(&tmp_dev->ref_count);
atomic_read       107 drivers/dma-buf/dma-fence-array.c 	return atomic_read(&array->num_pending) <= 0;
atomic_read      1085 drivers/dma/dmaengine.c 	if (atomic_read(idr_ref) == 0) {
atomic_read       238 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 		atomic_read(&chan->descs_allocated));
atomic_read       370 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 		 axi_chan_name(chan), atomic_read(&chan->descs_allocated));
atomic_read       429 drivers/dma/mediatek/mtk-hsdma.c 	reserved = min_t(u16, num_sgs, atomic_read(&pc->nr_free));
atomic_read       629 drivers/dma/mediatek/mtk-hsdma.c 	if (atomic_read(&pc->nr_free) >= MTK_DMA_SIZE - 1)
atomic_read        30 drivers/dma/qcom/hidma_dbg.c 	seq_printf(s, "allocated=%d\n", atomic_read(&tre->allocated));
atomic_read        70 drivers/dma/qcom/hidma_dbg.c 			atomic_read(&lldev->pending_tre_count));
atomic_read       126 drivers/dma/qcom/hidma_ll.c 	if (atomic_read(&tre->allocated) != true) {
atomic_read       306 drivers/dma/qcom/hidma_ll.c 	while (atomic_read(&lldev->pending_tre_count)) {
atomic_read       612 drivers/dma/qcom/hidma_ll.c 	if (atomic_read(&tre->allocated) != true) {
atomic_read       570 drivers/dma/sun6i-dma.c 		if (!atomic_read(&sdev->tasklet_shutdown))
atomic_read        61 drivers/edac/edac_pci_sysfs.c 	return sprintf(data, "%u\n", atomic_read(&pci->counters.pe_count));
atomic_read        67 drivers/edac/edac_pci_sysfs.c 	return sprintf(data, "%u\n", atomic_read(&pci->counters.npe_count));
atomic_read       649 drivers/edac/edac_pci_sysfs.c 	before_count = atomic_read(&pci_parity_count);
atomic_read       661 drivers/edac/edac_pci_sysfs.c 		if (before_count != atomic_read(&pci_parity_count))
atomic_read       430 drivers/edac/thunderx_edac.c 	while (!atomic_read(&lmc->ecc_int) && timeout--) {
atomic_read       314 drivers/firewire/core-card.c 			atomic_read(&root_device->state) == FW_DEVICE_RUNNING;
atomic_read      1001 drivers/firewire/core-device.c 		    atomic_read(&device->state) == FW_DEVICE_INITIALIZING) {
atomic_read      1179 drivers/firewire/core-device.c 	    atomic_read(&device->state) == FW_DEVICE_INITIALIZING) {
atomic_read      1283 drivers/firewire/core-device.c 		if (atomic_read(&device->state) == FW_DEVICE_RUNNING) {
atomic_read       144 drivers/firewire/nosy.c 				     atomic_read(&buffer->size) > 0) ||
atomic_read       148 drivers/firewire/nosy.c 	if (atomic_read(&buffer->size) == 0)
atomic_read       188 drivers/firewire/nosy.c 	    atomic_read(&buffer->size) + sizeof(struct packet) + length) {
atomic_read       326 drivers/firewire/nosy.c 	if (atomic_read(&client->buffer.size) > 0)
atomic_read       612 drivers/firmware/efi/efi.c 			for (i = 0; i < atomic_read(&rsv->count); i++) {
atomic_read      1101 drivers/gpio/gpio-pca953x.c 	if (atomic_read(&chip->wakeup_path))
atomic_read      1114 drivers/gpio/gpio-pca953x.c 	if (!atomic_read(&chip->wakeup_path)) {
atomic_read       550 drivers/gpio/gpio-rcar.c 	if (atomic_read(&p->wakeup_path))
atomic_read      1331 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		is_invalid_userptr = atomic_read(&mem->invalid);
atomic_read      1583 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (atomic_read(&adev->gmc.vm_fault_info_updated) == 1) {
atomic_read      1698 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if (!atomic_read(&mem->invalid))
atomic_read      1724 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		invalid = atomic_read(&mem->invalid);
atomic_read      1886 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	evicted_bos = atomic_read(&process_info->evicted_bos);
atomic_read      1909 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if (atomic_read(&process_info->evicted_bos) != evicted_bos)
atomic_read       132 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	if (atomic_read(&p->ctx->guilty) == 1) {
atomic_read       117 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	ctx->reset_counter = atomic_read(&adev->gpu_reset_counter);
atomic_read       119 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	ctx->vram_lost_counter = atomic_read(&adev->vram_lost_counter);
atomic_read       336 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	reset_counter = atomic_read(&adev->gpu_reset_counter);
atomic_read       370 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	if (ctx->reset_counter != atomic_read(&adev->gpu_reset_counter))
atomic_read       373 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	if (ctx->vram_lost_counter != atomic_read(&adev->vram_lost_counter))
atomic_read       376 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	if (atomic_read(&ctx->guilty))
atomic_read       938 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	last_seq = atomic_read(&ring->fence_drv.last_seq);
atomic_read      1003 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	if (preempt_seq <= atomic_read(&drv->last_seq))
atomic_read      1061 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	if (atomic_read(&ring->fence_drv.last_seq) !=
atomic_read      3938 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 			dev_info(tmp_adev->dev, "GPU reset(%d) failed\n", atomic_read(&adev->gpu_reset_counter));
atomic_read      3941 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 			dev_info(tmp_adev->dev, "GPU reset(%d) succeeded!\n", atomic_read(&adev->gpu_reset_counter));
atomic_read       122 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		seq = atomic_read(&drv->last_seq);
atomic_read       241 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		last_seq = atomic_read(&ring->fence_drv.last_seq);
atomic_read       372 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	emitted -= atomic_read(&ring->fence_drv.last_seq);
atomic_read       406 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	amdgpu_fence_write(ring, atomic_read(&ring->fence_drv.last_seq));
atomic_read       713 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 			   atomic_read(&ring->fence_drv.last_seq));
atomic_read       182 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		atomic_read(&adev->gpu_reset_counter);
atomic_read       565 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c 	return !!atomic_read(&src->enabled_types[type]);
atomic_read        47 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 		  job->base.sched->name, atomic_read(&ring->fence_drv.last_seq),
atomic_read        83 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 	(*job)->vram_lost_counter = atomic_read(&adev->vram_lost_counter);
atomic_read       230 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 	if (job->vram_lost_counter != atomic_read(&ring->adev->vram_lost_counter))
atomic_read       918 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		ui32 = atomic_read(&adev->vram_lost_counter);
atomic_read       187 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c 				|| atomic_read(&ring->num_jobs[i])) {
atomic_read       355 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			if (atomic_read(&adev->uvd.handles[i]))
atomic_read       425 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		uint32_t handle = atomic_read(&adev->uvd.handles[i]);
atomic_read       747 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			if (atomic_read(&adev->uvd.handles[i]) == handle) {
atomic_read       771 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			if (atomic_read(&adev->uvd.handles[i]) == handle) {
atomic_read      1287 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		if (atomic_read(&adev->uvd.handles[i]))
atomic_read       267 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 		if (atomic_read(&adev->vce.handles[i]))
atomic_read       407 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 		uint32_t handle = atomic_read(&adev->vce.handles[i]);
atomic_read       680 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 		if (atomic_read(&p->adev->vce.handles[i]) == handle) {
atomic_read      1088 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			atomic_read(&adev->gpu_reset_counter);
atomic_read      1783 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	enable = !!atomic_read(&adev->vm_manager.num_prt_users);
atomic_read      1275 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c 		&& !atomic_read(&adev->gmc.vm_fault_info_updated)) {
atomic_read      1460 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c 		&& !atomic_read(&adev->gmc.vm_fault_info_updated)) {
atomic_read       788 drivers/gpu/drm/amd/amdkfd/kfd_device.c 	return  (atomic_read(&kfd_locked) > 0);
atomic_read      1019 drivers/gpu/drm/amd/amdkfd/kfd_events.c 	int reset_cause = atomic_read(&dev->sram_ecc_flag) ?
atomic_read       202 drivers/gpu/drm/arm/hdlcd_drv.c 	seq_printf(m, "underrun : %d\n", atomic_read(&hdlcd->buffer_underrun_count));
atomic_read       203 drivers/gpu/drm/arm/hdlcd_drv.c 	seq_printf(m, "dma_end  : %d\n", atomic_read(&hdlcd->dma_end_count));
atomic_read       204 drivers/gpu/drm/arm/hdlcd_drv.c 	seq_printf(m, "bus_error: %d\n", atomic_read(&hdlcd->bus_error_count));
atomic_read       205 drivers/gpu/drm/arm/hdlcd_drv.c 	seq_printf(m, "vsync    : %d\n", atomic_read(&hdlcd->vsync_count));
atomic_read       183 drivers/gpu/drm/arm/malidp_drv.c 			atomic_read(&malidp->config_valid) == MALIDP_CONFIG_VALID_DONE,
atomic_read      1328 drivers/gpu/drm/arm/malidp_hw.c 			if ((atomic_read(&malidp->config_valid) != MALIDP_CONFIG_START) ||
atomic_read      1311 drivers/gpu/drm/drm_bufs.c 	if (atomic_read(&dev->buf_alloc)) {
atomic_read      1503 drivers/gpu/drm/drm_bufs.c 	if (atomic_read(&dev->buf_alloc)) {
atomic_read       294 drivers/gpu/drm/drm_dp_aux_dev.c 	wait_var_event(&aux_dev->usecount, !atomic_read(&aux_dev->usecount));
atomic_read       403 drivers/gpu/drm/drm_vblank.c 	if (atomic_read(&vblank->refcount) == 0 && vblank->enabled) {
atomic_read      1043 drivers/gpu/drm/drm_vblank.c 	if (WARN_ON(atomic_read(&vblank->refcount) == 0))
atomic_read      1291 drivers/gpu/drm/drm_vblank.c 	if (atomic_read(&vblank->refcount) != 0 || drm_vblank_offdelay == 0)
atomic_read      1787 drivers/gpu/drm/drm_vblank.c 		       !atomic_read(&vblank->refcount));
atomic_read        73 drivers/gpu/drm/etnaviv/etnaviv_gem.h 	return atomic_read(&etnaviv_obj->gpu_active) != 0;
atomic_read      1804 drivers/gpu/drm/etnaviv/etnaviv_gpu.c 	if (atomic_read(&gpu->sched.hw_rq_count))
atomic_read        98 drivers/gpu/drm/exynos/exynos7_drm_decon.c 				!atomic_read(&ctx->wait_vsync_event),
atomic_read       602 drivers/gpu/drm/exynos/exynos7_drm_decon.c 		if (atomic_read(&ctx->wait_vsync_event)) {
atomic_read       314 drivers/gpu/drm/exynos/exynos_drm_fimd.c 				!atomic_read(&ctx->wait_vsync_event),
atomic_read       953 drivers/gpu/drm/exynos/exynos_drm_fimd.c 	if (atomic_read(&ctx->triggering))
atomic_read       992 drivers/gpu/drm/exynos/exynos_drm_fimd.c 	if (atomic_read(&ctx->wait_vsync_event)) {
atomic_read      1045 drivers/gpu/drm/exynos/exynos_drm_fimd.c 		if (atomic_read(&ctx->wait_vsync_event)) {
atomic_read       391 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (atomic_read(&g2d_userptr->refcount) > 0)
atomic_read        79 drivers/gpu/drm/gma500/mmu.c 	if (atomic_read(&driver->needs_tlbflush) || force) {
atomic_read       110 drivers/gpu/drm/gma500/mmu.c 	if (atomic_read(&driver->needs_tlbflush))
atomic_read      4282 drivers/gpu/drm/i915/display/intel_display.c 	if (atomic_read(&dev_priv->gpu_error.pending_fb_pin)) {
atomic_read      5594 drivers/gpu/drm/i915/display/intel_dp.c #define C (hdcp->cp_irq_count_cached != atomic_read(&hdcp->cp_irq_count))
atomic_read      5998 drivers/gpu/drm/i915/display/intel_dp.c 	hdcp->cp_irq_count_cached = atomic_read(&hdcp->cp_irq_count);
atomic_read       300 drivers/gpu/drm/i915/display/intel_frontbuffer.c 		WARN_ON(!(atomic_read(&old->bits) & frontbuffer_bits));
atomic_read       305 drivers/gpu/drm/i915/display/intel_frontbuffer.c 		WARN_ON(atomic_read(&new->bits) & frontbuffer_bits);
atomic_read        83 drivers/gpu/drm/i915/display/intel_frontbuffer.h 	frontbuffer_bits = atomic_read(&front->bits);
atomic_read       111 drivers/gpu/drm/i915/display/intel_frontbuffer.h 	frontbuffer_bits = atomic_read(&front->bits);
atomic_read       202 drivers/gpu/drm/i915/gem/i915_gem_context.c 		if (atomic_read(&ctx->hw_id_pin_count)) {
atomic_read       648 drivers/gpu/drm/i915/gem/i915_gem_context.c 	GEM_BUG_ON(!atomic_read(&ctx->hw_id_pin_count));
atomic_read      1040 drivers/gpu/drm/i915/gem/i915_gem_context.c 		return !atomic_read(&ce->pin_count);
atomic_read      2073 drivers/gpu/drm/i915/gem/i915_gem_context.c 	return atomic_read(&file_priv->ban_score) >= I915_CLIENT_SCORE_BANNED;
atomic_read      2330 drivers/gpu/drm/i915/gem/i915_gem_context.c 	args->batch_active = atomic_read(&ctx->guilty_count);
atomic_read      2331 drivers/gpu/drm/i915/gem/i915_gem_context.c 	args->batch_pending = atomic_read(&ctx->active_count);
atomic_read      2349 drivers/gpu/drm/i915/gem/i915_gem_context.c 		GEM_BUG_ON(atomic_read(&ctx->hw_id_pin_count));
atomic_read      2358 drivers/gpu/drm/i915/gem/i915_gem_context.c 	GEM_BUG_ON(atomic_read(&ctx->hw_id_pin_count) == ~0u);
atomic_read       126 drivers/gpu/drm/i915/gem/i915_gem_context.h 	GEM_BUG_ON(atomic_read(&ctx->hw_id_pin_count) == 0u);
atomic_read       223 drivers/gpu/drm/i915/gem/i915_gem_domain.c 	if (atomic_read(&obj->bind_count)) {
atomic_read       124 drivers/gpu/drm/i915/gem/i915_gem_object.c 			GEM_BUG_ON(!atomic_read(&vma->open_count));
atomic_read       146 drivers/gpu/drm/i915/gem/i915_gem_object.c 	GEM_BUG_ON(!atomic_read(&i915->mm.free_count));
atomic_read       174 drivers/gpu/drm/i915/gem/i915_gem_object.c 		GEM_BUG_ON(atomic_read(&obj->bind_count));
atomic_read       270 drivers/gpu/drm/i915/gem/i915_gem_object.h 	return atomic_read(&obj->mm.pages_pin_count);
atomic_read       191 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	GEM_BUG_ON(atomic_read(&obj->bind_count));
atomic_read       195 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	if (unlikely(atomic_read(&obj->mm.pages_pin_count))) {
atomic_read        72 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	if (atomic_read(&obj->mm.pages_pin_count) > atomic_read(&obj->bind_count))
atomic_read       244 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 			    atomic_read(&obj->bind_count))
atomic_read        42 drivers/gpu/drm/i915/gem/selftests/i915_gem_phys.c 	if (!atomic_read(&obj->mm.pages_pin_count)) {
atomic_read        53 drivers/gpu/drm/i915/gt/intel_context.c 	if (likely(!atomic_read(&ce->pin_count))) {
atomic_read        54 drivers/gpu/drm/i915/gt/intel_context.h 	return atomic_read(&ce->pin_count);
atomic_read      1269 drivers/gpu/drm/i915/gt/intel_engine_cs.c 			   enableddisabled(!atomic_read(&engine->execlists.tasklet.count)));
atomic_read      1374 drivers/gpu/drm/i915/gt/intel_engine_cs.c 	drm_printf(m, "\tAwake? %d\n", atomic_read(&engine->wakeref.count));
atomic_read      2379 drivers/gpu/drm/i915/gt/intel_lrc.c 		  atomic_read(&execlists->tasklet.count));
atomic_read      2662 drivers/gpu/drm/i915/gt/intel_lrc.c 		  atomic_read(&execlists->tasklet.count));
atomic_read        74 drivers/gpu/drm/i915/gt/intel_reset.c 				 atomic_read(&file_priv->ban_score));
atomic_read       102 drivers/gpu/drm/i915/gt/intel_reset.c 				 ctx->name, atomic_read(&ctx->guilty_count));
atomic_read       283 drivers/gpu/drm/i915/gt/intel_timeline.c 	GEM_BUG_ON(atomic_read(&timeline->pin_count));
atomic_read       344 drivers/gpu/drm/i915/gt/intel_timeline.c 	GEM_BUG_ON(!atomic_read(&tl->pin_count));
atomic_read       379 drivers/gpu/drm/i915/gt/intel_timeline.c 	GEM_BUG_ON(!atomic_read(&tl->pin_count));
atomic_read       529 drivers/gpu/drm/i915/gt/intel_timeline.c 	GEM_BUG_ON(!atomic_read(&tl->pin_count));
atomic_read        42 drivers/gpu/drm/i915/gt/mock_engine.c 	GEM_BUG_ON(!atomic_read(&tl->pin_count));
atomic_read       748 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		  atomic_read(&execlists->tasklet.count));
atomic_read       924 drivers/gpu/drm/i915/gvt/gtt.c 	int v = atomic_read(&spt->refcount);
atomic_read       932 drivers/gpu/drm/i915/gvt/gtt.c 	int v = atomic_read(&spt->refcount);
atomic_read      1954 drivers/gpu/drm/i915/gvt/gtt.c 	if (GEM_WARN_ON(atomic_read(&mm->pincount)))
atomic_read      2024 drivers/gpu/drm/i915/gvt/gtt.c 		if (atomic_read(&mm->pincount))
atomic_read        68 drivers/gpu/drm/i915/gvt/scheduler.c 	if (WARN_ON(!atomic_read(&workload->shadow_mm->pincount)))
atomic_read       923 drivers/gpu/drm/i915/gvt/scheduler.c 			   !atomic_read(&workload->shadow_ctx_active));
atomic_read      1076 drivers/gpu/drm/i915/gvt/scheduler.c 	if (atomic_read(&s->running_workload_num)) {
atomic_read      1080 drivers/gpu/drm/i915/gvt/scheduler.c 				!atomic_read(&s->running_workload_num));
atomic_read       234 drivers/gpu/drm/i915/gvt/vgpu.c 	if (atomic_read(&vgpu->submission.running_workload_num)) {
atomic_read       159 drivers/gpu/drm/i915/i915_active.c 	GEM_BUG_ON(!atomic_read(&ref->count));
atomic_read       340 drivers/gpu/drm/i915/i915_active.c 	GEM_BUG_ON(!atomic_read(&ref->count));
atomic_read       360 drivers/gpu/drm/i915/i915_active.c 	if (!atomic_read(&ref->count) && ref->active)
atomic_read       490 drivers/gpu/drm/i915/i915_active.c 	GEM_BUG_ON(atomic_read(&ref->count));
atomic_read       393 drivers/gpu/drm/i915/i915_active.h 	return !atomic_read(&ref->count);
atomic_read       248 drivers/gpu/drm/i915/i915_debugfs.c 	if (!atomic_read(&obj->bind_count))
atomic_read       361 drivers/gpu/drm/i915/i915_debugfs.c 		   atomic_read(&i915->mm.free_count),
atomic_read       659 drivers/gpu/drm/i915/i915_debugfs.c 			   i, atomic_read(&reg->pin_count));
atomic_read      1583 drivers/gpu/drm/i915/i915_debugfs.c 				   atomic_read(&ctx->hw_id_pin_count));
atomic_read      1734 drivers/gpu/drm/i915/i915_debugfs.c 		   atomic_read(&rps->num_waiters));
atomic_read      1749 drivers/gpu/drm/i915/i915_debugfs.c 	seq_printf(m, "Wait boosts: %d\n", atomic_read(&rps->boosts));
atomic_read      2327 drivers/gpu/drm/i915/i915_debugfs.c 		   atomic_read(&dev_priv->drm.dev->power.usage_count));
atomic_read      2791 drivers/gpu/drm/i915/i915_debugfs.c 		   atomic_read(&dev_priv->gt.wakeref.count));
atomic_read      2690 drivers/gpu/drm/i915/i915_drv.c 	WARN_ON_ONCE(atomic_read(&rpm->wakeref_count));
atomic_read      2259 drivers/gpu/drm/i915/i915_drv.h 	while (atomic_read(&i915->mm.free_count)) {
atomic_read      2325 drivers/gpu/drm/i915/i915_drv.h 	return atomic_read(&error->reset_count);
atomic_read      2331 drivers/gpu/drm/i915/i915_drv.h 	return atomic_read(&error->reset_engine_count[engine->uabi_class]);
atomic_read      1683 drivers/gpu/drm/i915/i915_gem.c 	GEM_BUG_ON(atomic_read(&dev_priv->mm.free_count));
atomic_read        94 drivers/gpu/drm/i915/i915_gem.h 	return !atomic_read(&t->count);
atomic_read       308 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	if (atomic_read(&fence->pin_count))
atomic_read       321 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		if (atomic_read(&fence->pin_count))
atomic_read       355 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		GEM_BUG_ON(atomic_read(&fence->pin_count));
atomic_read       435 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		count += !atomic_read(&fence->pin_count);
atomic_read       776 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(atomic_read(px_used(pd)) > 2 * ARRAY_SIZE(pd->entry));
atomic_read       791 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(atomic_read(px_used(pd)) == 0);
atomic_read       975 drivers/gpu/drm/i915/i915_gem_gtt.c 	    idx, len, atomic_read(px_used(pd)));
atomic_read       976 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(!len || len >= atomic_read(px_used(pd)));
atomic_read      1002 drivers/gpu/drm/i915/i915_gem_gtt.c 			    atomic_read(&pt->used));
atomic_read      1003 drivers/gpu/drm/i915/i915_gem_gtt.c 			GEM_BUG_ON(!count || count >= atomic_read(&pt->used));
atomic_read      1051 drivers/gpu/drm/i915/i915_gem_gtt.c 	    idx, len, atomic_read(px_used(pd)));
atomic_read      1055 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(!atomic_read(px_used(pd))); /* Must be pinned! */
atomic_read      1111 drivers/gpu/drm/i915/i915_gem_gtt.c 			GEM_BUG_ON(!atomic_read(&pt->used));
atomic_read      1118 drivers/gpu/drm/i915/i915_gem_gtt.c 			    atomic_read(&pt->used));
atomic_read      1122 drivers/gpu/drm/i915/i915_gem_gtt.c 			GEM_BUG_ON(atomic_read(&pt->used) > 2 * I915_PDES);
atomic_read      1624 drivers/gpu/drm/i915/i915_gem_gtt.c 		GEM_BUG_ON(count > atomic_read(&pt->used));
atomic_read      1852 drivers/gpu/drm/i915/i915_gem_gtt.c 		if (px_base(pt) == scratch || atomic_read(&pt->used))
atomic_read        43 drivers/gpu/drm/i915/i915_globals.c 	if (park.epoch == atomic_read(&epoch))
atomic_read      1268 drivers/gpu/drm/i915/i915_gpu_error.c 	e->guilty = atomic_read(&ctx->guilty_count);
atomic_read      1269 drivers/gpu/drm/i915/i915_gpu_error.c 	e->active = atomic_read(&ctx->active_count);
atomic_read      1626 drivers/gpu/drm/i915/i915_gpu_error.c 	error->wakelock = atomic_read(&i915->runtime_pm.wakeref_count);
atomic_read      1182 drivers/gpu/drm/i915/i915_irq.c 		client_boost = atomic_read(&rps->num_waiters);
atomic_read       292 drivers/gpu/drm/i915/i915_request.c 		GEM_BUG_ON(!atomic_read(&rq->i915->gt_pm.rps.num_waiters));
atomic_read        97 drivers/gpu/drm/i915/i915_sw_fence.h 	return atomic_read(&fence->pending) <= 0;
atomic_read       102 drivers/gpu/drm/i915/i915_sw_fence.h 	return atomic_read(&fence->pending) < 0;
atomic_read       325 drivers/gpu/drm/i915/i915_sysfs.c 		boost = atomic_read(&rps->num_waiters);
atomic_read       524 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(atomic_read(&obj->mm.pages_pin_count) < atomic_read(&obj->bind_count));
atomic_read       429 drivers/gpu/drm/i915/i915_vma.h 	GEM_BUG_ON(atomic_read(&vma->fence->pin_count) <= 0);
atomic_read       140 drivers/gpu/drm/i915/intel_runtime_pm.c 		 rpm->debug.count, atomic_read(&rpm->wakeref_count))) {
atomic_read       598 drivers/gpu/drm/i915/intel_runtime_pm.c 	int count = atomic_read(&rpm->wakeref_count);
atomic_read       121 drivers/gpu/drm/i915/intel_runtime_pm.h 	__assert_rpm_raw_wakeref_held(rpm, atomic_read(&rpm->wakeref_count));
atomic_read       127 drivers/gpu/drm/i915/intel_runtime_pm.h 	__assert_rpm_wakelock_held(rpm, atomic_read(&rpm->wakeref_count));
atomic_read        34 drivers/gpu/drm/i915/intel_wakeref.c 	if (!atomic_read(&wf->count)) {
atomic_read        51 drivers/gpu/drm/i915/intel_wakeref.c 	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
atomic_read       121 drivers/gpu/drm/i915/intel_wakeref.h 	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
atomic_read       173 drivers/gpu/drm/i915/intel_wakeref.h 	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count));
atomic_read       128 drivers/gpu/drm/i915/selftests/i915_active.c 	if (atomic_read(&active->base.count) != count) {
atomic_read       130 drivers/gpu/drm/i915/selftests/i915_active.c 		       atomic_read(&active->base.count), count);
atomic_read        80 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		if (atomic_read(&obj->bind_count))
atomic_read       384 drivers/gpu/drm/i915/selftests/i915_request.c 			       atomic_read(&wait->pending), count,
atomic_read        44 drivers/gpu/drm/mga/mga_irq.c 	return atomic_read(&dev_priv->vbl_received);
atomic_read       129 drivers/gpu/drm/mga/mga_irq.c 		    (((cur_fence = atomic_read(&dev_priv->last_fence_retired))
atomic_read       155 drivers/gpu/drm/msm/adreno/a5xx_gpu.h 	int preempt_state = atomic_read(&a5xx_gpu->preempt_state);
atomic_read        31 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 			atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx]));
atomic_read        88 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	enable_count = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx]);
atomic_read       122 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	counts = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idxs[0]]);
atomic_read       151 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	enable_count = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx]);
atomic_read       177 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	counts = atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idxs[0]]);
atomic_read       268 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 			atomic_read(&dpu_kms->irq_obj.enable_counts[irq_idx]))
atomic_read       320 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 		irq_count = atomic_read(&irq_obj->irq_counts[i]);
atomic_read       321 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 		enable_count = atomic_read(&irq_obj->enable_counts[i]);
atomic_read       396 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 		if (atomic_read(&dpu_kms->irq_obj.enable_counts[i]) ||
atomic_read       315 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c 		if (atomic_read(&dpu_crtc->frame_pending) < 1) {
atomic_read       588 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c 	if (!atomic_read(&dpu_crtc->frame_pending)) {
atomic_read       726 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c 				atomic_read(&dpu_crtc->frame_pending));
atomic_read       731 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c 	if (atomic_read(&dpu_crtc->frame_pending)) {
atomic_read       733 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c 				     atomic_read(&dpu_crtc->frame_pending));
atomic_read       222 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.h 	return crtc ? atomic_read(&to_dpu_crtc(crtc)->frame_pending) : -EINVAL;
atomic_read       262 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 		      atomic_read(wait_info->atomic_cnt));
atomic_read       280 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				      atomic_read(wait_info->atomic_cnt));
atomic_read       292 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				      atomic_read(wait_info->atomic_cnt));
atomic_read       299 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 			atomic_read(wait_info->atomic_cnt));
atomic_read      1291 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				  atomic_read(&phy_enc->underrun_cnt));
atomic_read      1506 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				atomic_read(info->atomic_cnt) == 0, jiffies);
atomic_read      1511 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 						 atomic_read(info->atomic_cnt));
atomic_read      1513 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 	} while (atomic_read(info->atomic_cnt) && (rc == 0) &&
atomic_read      1888 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				atomic_read(&phys->vsync_cnt),
atomic_read      1889 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				atomic_read(&phys->underrun_cnt));
atomic_read       217 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 		     atomic_read(&phys_enc->pending_kickoff_cnt),
atomic_read       227 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 			  atomic_read(&phys_enc->pending_kickoff_cnt));
atomic_read       283 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 	refcount = atomic_read(&phys_enc->vblank_refcount);
atomic_read       328 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 			enable, atomic_read(&phys_enc->vblank_refcount));
atomic_read       601 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 		      atomic_read(&phys_enc->pending_kickoff_cnt));
atomic_read       618 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 			atomic_read(&phys_enc->pending_kickoff_cnt));
atomic_read       309 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 	old_cnt  = atomic_read(&phys_enc->pending_kickoff_cnt);
atomic_read       403 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 	refcount = atomic_read(&phys_enc->vblank_refcount);
atomic_read       416 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 		      atomic_read(&phys_enc->vblank_refcount));
atomic_read       651 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 			    atomic_read(&phys_enc->vblank_refcount));
atomic_read        54 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c 	if (atomic_read(&hw_blk->refcount))
atomic_read        81 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c 						atomic_read(&curr->refcount)))
atomic_read       127 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c 			atomic_read(&hw_blk->refcount));
atomic_read       129 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_blk.c 	if (!atomic_read(&hw_blk->refcount)) {
atomic_read        64 drivers/gpu/drm/nouveau/nouveau_chan.c 	if (likely(chan && chan->fence && !atomic_read(&chan->killed))) {
atomic_read       502 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c 	if (retrain || !atomic_read(&dp->lt.done))
atomic_read       546 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c 		if (atomic_read(&dp->lt.done))
atomic_read        53 drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c 		args->v0.ref = atomic_read(&chan->ref);
atomic_read       340 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c 		wait_event(clk->wait, !atomic_read(&clk->waiting));
atomic_read       232 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR]))
atomic_read       470 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR]))
atomic_read       189 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 		if (!atomic_read(&vmm->engref[i]))
atomic_read        34 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c 	if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR]))
atomic_read        44 drivers/gpu/drm/panfrost/panfrost_gem_shrinker.c 	if (atomic_read(&bo->gpu_usecount))
atomic_read       591 drivers/gpu/drm/panfrost/panfrost_job.c 		if (atomic_read(&js->queue[i].sched.hw_rq_count))
atomic_read       171 drivers/gpu/drm/panfrost/panfrost_mmu.c 			if (!atomic_read(&lru_mmu->as_count))
atomic_read       201 drivers/gpu/drm/panfrost/panfrost_mmu.c 	WARN_ON(atomic_read(&mmu->as_count) < 0);
atomic_read       294 drivers/gpu/drm/qxl/qxl_cmd.c 	irq_num = atomic_read(&qdev->irq_received_io_cmd);
atomic_read       298 drivers/gpu/drm/qxl/qxl_cmd.c 							       atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ);
atomic_read       301 drivers/gpu/drm/qxl/qxl_cmd.c 						 atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ);
atomic_read       305 drivers/gpu/drm/qxl/qxl_cmd.c 		irq_num = atomic_read(&qdev->irq_received_io_cmd);
atomic_read       311 drivers/gpu/drm/qxl/qxl_cmd.c 						       atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ);
atomic_read       314 drivers/gpu/drm/qxl/qxl_cmd.c 					 atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ);
atomic_read        44 drivers/gpu/drm/qxl/qxl_debugfs.c 	seq_printf(m, "%d\n", atomic_read(&qdev->irq_received));
atomic_read        45 drivers/gpu/drm/qxl/qxl_debugfs.c 	seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_display));
atomic_read        46 drivers/gpu/drm/qxl/qxl_debugfs.c 	seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_cursor));
atomic_read        47 drivers/gpu/drm/qxl/qxl_debugfs.c 	seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_io_cmd));
atomic_read        47 drivers/gpu/drm/r128/r128_irq.c 	return atomic_read(&dev_priv->vbl_received);
atomic_read       330 drivers/gpu/drm/r128/r128_state.c 	if (atomic_read(&dev_priv->idle_count) == 0)
atomic_read      7076 drivers/gpu/drm/radeon/cik.c 	if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read      7080 drivers/gpu/drm/radeon/cik.c 	if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) {
atomic_read      7123 drivers/gpu/drm/radeon/cik.c 	if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) {
atomic_read      7167 drivers/gpu/drm/radeon/cik.c 	if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) {
atomic_read      7172 drivers/gpu/drm/radeon/cik.c 	if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) {
atomic_read      7178 drivers/gpu/drm/radeon/cik.c 	    atomic_read(&rdev->irq.pflip[0])) {
atomic_read      7183 drivers/gpu/drm/radeon/cik.c 	    atomic_read(&rdev->irq.pflip[1])) {
atomic_read      7188 drivers/gpu/drm/radeon/cik.c 	    atomic_read(&rdev->irq.pflip[2])) {
atomic_read      7193 drivers/gpu/drm/radeon/cik.c 	    atomic_read(&rdev->irq.pflip[3])) {
atomic_read      7198 drivers/gpu/drm/radeon/cik.c 	    atomic_read(&rdev->irq.pflip[4])) {
atomic_read      7203 drivers/gpu/drm/radeon/cik.c 	    atomic_read(&rdev->irq.pflip[5])) {
atomic_read      7607 drivers/gpu/drm/radeon/cik.c 				if (atomic_read(&rdev->irq.pflip[0]))
atomic_read      7637 drivers/gpu/drm/radeon/cik.c 				if (atomic_read(&rdev->irq.pflip[1]))
atomic_read      7667 drivers/gpu/drm/radeon/cik.c 				if (atomic_read(&rdev->irq.pflip[2]))
atomic_read      7697 drivers/gpu/drm/radeon/cik.c 				if (atomic_read(&rdev->irq.pflip[3]))
atomic_read      7727 drivers/gpu/drm/radeon/cik.c 				if (atomic_read(&rdev->irq.pflip[4]))
atomic_read      7757 drivers/gpu/drm/radeon/cik.c 				if (atomic_read(&rdev->irq.pflip[5]))
atomic_read      4523 drivers/gpu/drm/radeon/evergreen.c 		if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read      4527 drivers/gpu/drm/radeon/evergreen.c 		if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) {
atomic_read      4531 drivers/gpu/drm/radeon/evergreen.c 		if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) {
atomic_read      4536 drivers/gpu/drm/radeon/evergreen.c 		if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read      4543 drivers/gpu/drm/radeon/evergreen.c 	if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) {
atomic_read      4550 drivers/gpu/drm/radeon/evergreen.c 		if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) {
atomic_read      4580 drivers/gpu/drm/radeon/evergreen.c 		    atomic_read(&rdev->irq.pflip[i]), "vblank", i);
atomic_read      4761 drivers/gpu/drm/radeon/evergreen.c 				if (atomic_read(&rdev->irq.pflip[crtc_idx])) {
atomic_read       721 drivers/gpu/drm/radeon/r100.c 	if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read       725 drivers/gpu/drm/radeon/r100.c 	    atomic_read(&rdev->irq.pflip[0])) {
atomic_read       729 drivers/gpu/drm/radeon/r100.c 	    atomic_read(&rdev->irq.pflip[1])) {
atomic_read       794 drivers/gpu/drm/radeon/r100.c 			if (atomic_read(&rdev->irq.pflip[0]))
atomic_read       803 drivers/gpu/drm/radeon/r100.c 			if (atomic_read(&rdev->irq.pflip[1]))
atomic_read      3822 drivers/gpu/drm/radeon/r600.c 	if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read      3828 drivers/gpu/drm/radeon/r600.c 	if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) {
atomic_read      3834 drivers/gpu/drm/radeon/r600.c 	    atomic_read(&rdev->irq.pflip[0])) {
atomic_read      3839 drivers/gpu/drm/radeon/r600.c 	    atomic_read(&rdev->irq.pflip[1])) {
atomic_read      4146 drivers/gpu/drm/radeon/r600.c 				if (atomic_read(&rdev->irq.pflip[0]))
atomic_read      4176 drivers/gpu/drm/radeon/r600.c 				if (atomic_read(&rdev->irq.pflip[1]))
atomic_read       594 drivers/gpu/drm/radeon/radeon_kms.c 		*value = atomic_read(&rdev->gpu_reset_counter);
atomic_read       257 drivers/gpu/drm/radeon/radeon_ring.c 	if (rptr != atomic_read(&ring->last_rptr)) {
atomic_read       257 drivers/gpu/drm/radeon/radeon_uvd.c 		uint32_t handle = atomic_read(&rdev->uvd.handles[i]);
atomic_read       332 drivers/gpu/drm/radeon/radeon_uvd.c 		uint32_t handle = atomic_read(&rdev->uvd.handles[i]);
atomic_read       517 drivers/gpu/drm/radeon/radeon_uvd.c 			if (atomic_read(&p->rdev->uvd.handles[i]) == handle) {
atomic_read       543 drivers/gpu/drm/radeon/radeon_uvd.c 			if (atomic_read(&p->rdev->uvd.handles[i]) == handle) {
atomic_read       860 drivers/gpu/drm/radeon/radeon_uvd.c 		if (!atomic_read(&rdev->uvd.handles[i]))
atomic_read       205 drivers/gpu/drm/radeon/radeon_vce.c 		if (atomic_read(&rdev->vce.handles[i]))
atomic_read       320 drivers/gpu/drm/radeon/radeon_vce.c 		uint32_t handle = atomic_read(&rdev->vce.handles[i]);
atomic_read       528 drivers/gpu/drm/radeon/radeon_vce.c 		if (atomic_read(&p->rdev->vce.handles[i]) == handle) {
atomic_read       680 drivers/gpu/drm/radeon/rs600.c 	if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read       684 drivers/gpu/drm/radeon/rs600.c 	    atomic_read(&rdev->irq.pflip[0])) {
atomic_read       688 drivers/gpu/drm/radeon/rs600.c 	    atomic_read(&rdev->irq.pflip[1])) {
atomic_read       798 drivers/gpu/drm/radeon/rs600.c 			if (atomic_read(&rdev->irq.pflip[0]))
atomic_read       807 drivers/gpu/drm/radeon/rs600.c 			if (atomic_read(&rdev->irq.pflip[1]))
atomic_read      6081 drivers/gpu/drm/radeon/si.c 	if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
atomic_read      6085 drivers/gpu/drm/radeon/si.c 	if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) {
atomic_read      6089 drivers/gpu/drm/radeon/si.c 	if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) {
atomic_read      6093 drivers/gpu/drm/radeon/si.c 	if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) {
atomic_read      6098 drivers/gpu/drm/radeon/si.c 	if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) {
atomic_read      6121 drivers/gpu/drm/radeon/si.c 		    atomic_read(&rdev->irq.pflip[i]), "vblank", i);
atomic_read      6304 drivers/gpu/drm/radeon/si.c 				if (atomic_read(&rdev->irq.pflip[crtc_idx])) {
atomic_read        53 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->hw_job_count = atomic_read(
atomic_read       147 drivers/gpu/drm/scheduler/sched_entity.c 		num_jobs = atomic_read(&sched->num_jobs);
atomic_read       466 drivers/gpu/drm/scheduler/sched_entity.c 	if (entity->guilty && atomic_read(entity->guilty))
atomic_read       338 drivers/gpu/drm/scheduler/sched_main.c 					if (atomic_read(&bad->karma) >
atomic_read       487 drivers/gpu/drm/scheduler/sched_main.c 		if (!found_guilty && atomic_read(&s_job->karma) > sched->hang_limit) {
atomic_read       569 drivers/gpu/drm/scheduler/sched_main.c 	return atomic_read(&sched->hw_rq_count) <
atomic_read       123 drivers/gpu/drm/ttm/ttm_bo.c 				atomic_read(&glob->bo_count));
atomic_read       156 drivers/gpu/drm/ttm/ttm_bo.c 	BUG_ON(atomic_read(&bo->cpu_writers));
atomic_read       116 drivers/gpu/drm/ttm/ttm_execbuf_util.c 		if (!ret && unlikely(atomic_read(&bo->cpu_writers) > 0)) {
atomic_read        95 drivers/gpu/drm/ttm/ttm_module.c 	wait_event(exit_q, atomic_read(&device_released) == 1);
atomic_read        90 drivers/gpu/drm/vc4/vc4_hvs.c 	drm_printf(&p, "%d\n", atomic_read(&vc4->underrun));
atomic_read       100 drivers/gpu/drm/via/via_irq.c 	return atomic_read(&dev_priv->vbl_received);
atomic_read       116 drivers/gpu/drm/via/via_irq.c 		if (!(atomic_read(&dev_priv->vbl_received) & 0x0F)) {
atomic_read       126 drivers/gpu/drm/via/via_irq.c 		if (!(atomic_read(&dev_priv->vbl_received) & 0xFF)) {
atomic_read       241 drivers/gpu/drm/via/via_irq.c 		cur_irq_sequence = atomic_read(&cur_irq->irq_received);
atomic_read       245 drivers/gpu/drm/via/via_irq.c 			       atomic_read(&cur_irq->irq_received)) -
atomic_read       365 drivers/gpu/drm/via/via_irq.c 			atomic_read(&cur_irq->irq_received);
atomic_read       541 drivers/gpu/drm/virtio/virtgpu_ioctl.c 				 atomic_read(&cache_ent->is_valid), 5 * HZ);
atomic_read      1312 drivers/gpu/drm/vmwgfx/vmwgfx_drv.c 	if (atomic_read(&dev_priv->num_fifo_resources) != 0) {
atomic_read       547 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 		*seqno = atomic_read(&dev_priv->marker_seq);
atomic_read       153 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	ret = ((atomic_read(&dev_priv->marker_seq) - seqno)
atomic_read       192 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	signal_seq = atomic_read(&dev_priv->marker_seq);
atomic_read       138 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 			seqno = atomic_read(&dev_priv->marker_seq);
atomic_read       314 drivers/gpu/host1x/syncpt.c 	current_val = (u32)atomic_read(&sp->min_val);
atomic_read       315 drivers/gpu/host1x/syncpt.c 	future_val = (u32)atomic_read(&sp->max_val);
atomic_read       478 drivers/gpu/host1x/syncpt.c 	return (u32)atomic_read(&sp->max_val);
atomic_read       493 drivers/gpu/host1x/syncpt.c 	return (u32)atomic_read(&sp->min_val);
atomic_read        87 drivers/gpu/host1x/syncpt.h 	min = atomic_read(&sp->min_val);
atomic_read        88 drivers/gpu/host1x/syncpt.h 	max = atomic_read(&sp->max_val);
atomic_read       112 drivers/greybus/greybus_trace.h 		__entry->waiters = atomic_read(&operation->waiters);
atomic_read        90 drivers/greybus/operation.c 		if (atomic_read(&operation->waiters))
atomic_read       370 drivers/hid/hid-cp2112.c 		atomic_read(avail), msecs_to_jiffies(RESPONSE_TIMEOUT));
atomic_read      1838 drivers/hid/hid-logitech-hidpp.c 	s = atomic_read(&data->workqueue_size);
atomic_read      3354 drivers/hid/hid-logitech-hidpp.c 	bool connected = atomic_read(&hidpp->connected);
atomic_read       292 drivers/hsi/clients/ssi_protocol.c 		if (atomic_read(&ssi->tx_usecnt))
atomic_read       346 drivers/hsi/clients/ssi_protocol.c 	dev_dbg(&master->device, "start TX %d\n", atomic_read(&ssi->tx_usecnt));
atomic_read       363 drivers/hsi/clients/ssi_protocol.c 	WARN_ON_ONCE(atomic_read(&ssi->tx_usecnt) == 0);
atomic_read       374 drivers/hsi/clients/ssi_protocol.c 	dev_dbg(&master->device, "stop TX %d\n", atomic_read(&ssi->tx_usecnt));
atomic_read       467 drivers/hsi/clients/ssi_protocol.c 			if (atomic_read(&ssi->tx_usecnt) == 0)
atomic_read       863 drivers/hsi/clients/ssi_protocol.c 		if (atomic_read(&ssi->tx_usecnt)) {
atomic_read      1038 drivers/hv/channel_mgmt.c 	while (atomic_read(&vmbus_connection.offer_in_progress) != 0) {
atomic_read      1172 drivers/hv/hv_balloon.c 	if (status.hdr.trans_id != atomic_read(&trans_id))
atomic_read      2182 drivers/hv/vmbus_drv.c 	while (atomic_read(&vmbus_connection.offer_in_progress) != 0) {
atomic_read      2213 drivers/hv/vmbus_drv.c 	if (atomic_read(&vmbus_connection.nr_chan_close_on_suspend) > 0)
atomic_read      2216 drivers/hv/vmbus_drv.c 	WARN_ON(atomic_read(&vmbus_connection.nr_chan_fixup_on_resume) != 0);
atomic_read      2286 drivers/hv/vmbus_drv.c 	WARN_ON(atomic_read(&vmbus_connection.nr_chan_fixup_on_resume) == 0);
atomic_read        60 drivers/hwmon/pwm-fan.c 	pulses = atomic_read(&ctx->pulses);
atomic_read       539 drivers/hwmon/sht15.c 		if (!atomic_read(&data->interrupt_handled))
atomic_read       831 drivers/hwmon/sht15.c 		    || atomic_read(&data->interrupt_handled))
atomic_read       446 drivers/hwtracing/coresight/coresight-etb10.c 	if (atomic_read(csdev->refcnt) != 1)
atomic_read        85 drivers/hwtracing/coresight/coresight-funnel.c 	if (atomic_read(&csdev->refcnt[inport]) == 0) {
atomic_read       106 drivers/hwtracing/coresight/coresight-replicator.c 	if (atomic_read(&csdev->refcnt[outport]) == 0) {
atomic_read       348 drivers/hwtracing/coresight/coresight-tmc-etf.c 	if (atomic_read(&csdev->refcnt[0]) == 0) {
atomic_read       464 drivers/hwtracing/coresight/coresight-tmc-etf.c 	if (atomic_read(csdev->refcnt) != 1)
atomic_read      1489 drivers/hwtracing/coresight/coresight-tmc-etr.c 	if (atomic_read(csdev->refcnt) != 1) {
atomic_read       307 drivers/hwtracing/coresight/coresight.c 		if (atomic_read(&csdev->refcnt[i]) != 0)
atomic_read      1287 drivers/hwtracing/intel_th/msu.c 	if (atomic_read(&msc->user_count) != -1)
atomic_read      2570 drivers/i3c/master.c 	if (atomic_read(&dev->ibi->pending_ibis))
atomic_read       239 drivers/iio/accel/hid-sensor-accel-3d.c 	if (atomic_read(&accel_state->common_attributes.data_ready)) {
atomic_read       169 drivers/iio/adc/stm32-dfsdm-core.c 		atomic_read(&priv->n_active_ch));
atomic_read       211 drivers/iio/adc/stm32-dfsdm-core.c 		atomic_read(&priv->n_active_ch));
atomic_read       358 drivers/iio/adc/sun4i-gpadc-iio.c 	if (atomic_read(&info->ignore_temp_data_irq))
atomic_read       372 drivers/iio/adc/sun4i-gpadc-iio.c 	if (atomic_read(&info->ignore_fifo_data_irq))
atomic_read       215 drivers/iio/common/hid-sensors/hid-sensor-trigger.c 	if (atomic_read(&attrb->user_requested_state))
atomic_read       227 drivers/iio/common/hid-sensors/hid-sensor-trigger.c 	if (atomic_read(&attrb->runtime_pm_enable))
atomic_read       633 drivers/iio/common/ssp_sensors/ssp_dev.c 	if (atomic_read(&data->enable_refcount) > 0)
atomic_read       658 drivers/iio/common/ssp_sensors/ssp_dev.c 	if (atomic_read(&data->enable_refcount) > 0)
atomic_read       200 drivers/iio/gyro/hid-sensor-gyro-3d.c 	if (atomic_read(&gyro_state->common_attributes.data_ready))
atomic_read       129 drivers/iio/humidity/hid-sensor-humidity.c 	if (atomic_read(&humid_st->common_attributes.data_ready))
atomic_read       166 drivers/iio/industrialio-trigger.c 	if (!atomic_read(&trig->use_count)) {
atomic_read       190 drivers/iio/industrialio-trigger.c 	if (!atomic_read(&trig->use_count)) {
atomic_read       190 drivers/iio/light/hid-sensor-als.c 	if (atomic_read(&als_state->common_attributes.data_ready))
atomic_read       170 drivers/iio/light/hid-sensor-prox.c 	if (atomic_read(&prox_state->common_attributes.data_ready))
atomic_read       110 drivers/iio/light/lm3533-als.c 		*zone = atomic_read(&als->zone);
atomic_read       294 drivers/iio/magnetometer/hid-sensor-magn-3d.c 	if (atomic_read(&magn_state->magn_flux_attributes.data_ready))
atomic_read       199 drivers/iio/orientation/hid-sensor-incl-3d.c 	if (atomic_read(&incl_state->common_attributes.data_ready))
atomic_read       155 drivers/iio/orientation/hid-sensor-rotation.c 	if (atomic_read(&rot_state->common_attributes.data_ready))
atomic_read       174 drivers/iio/pressure/hid-sensor-press.c 	if (atomic_read(&press_state->common_attributes.data_ready))
atomic_read       127 drivers/iio/temperature/hid-sensor-temperature.c 	if (atomic_read(&temp_st->common_attributes.data_ready))
atomic_read       290 drivers/infiniband/core/cq.c 	if (WARN_ON_ONCE(atomic_read(&cq->usecnt)))
atomic_read       182 drivers/infiniband/core/fmr_pool.c 	if (atomic_read(&pool->flush_ser) - atomic_read(&pool->req_ser) < 0)
atomic_read       378 drivers/infiniband/core/fmr_pool.c 				     atomic_read(&pool->flush_ser) - serial >= 0))
atomic_read       214 drivers/infiniband/core/iwcm.c 	BUG_ON(atomic_read(&cm_id_priv->refcount)==0);
atomic_read        90 drivers/infiniband/core/iwpm_msg.c 	msg_seq = atomic_read(&echo_nlmsg_seq);
atomic_read       177 drivers/infiniband/core/iwpm_msg.c 	msg_seq = atomic_read(&echo_nlmsg_seq);
atomic_read       266 drivers/infiniband/core/iwpm_msg.c 	msg_seq = atomic_read(&echo_nlmsg_seq);
atomic_read       350 drivers/infiniband/core/iwpm_msg.c 	msg_seq = atomic_read(&echo_nlmsg_seq);
atomic_read        64 drivers/infiniband/core/iwpm_util.c 	if (atomic_read(&iwpm_admin.refcount) == 0) {
atomic_read       108 drivers/infiniband/core/iwpm_util.c 	if (atomic_read(&iwpm_admin.refcount) == 0) {
atomic_read      2311 drivers/infiniband/core/mad.c 	if (atomic_read(&qp_info->snoop_count))
atomic_read      2560 drivers/infiniband/core/mad.c 	if (atomic_read(&qp_info->snoop_count))
atomic_read      2804 drivers/infiniband/core/mad.c 			if (atomic_read(&recv_mad_agent->qp_info->snoop_count))
atomic_read      2822 drivers/infiniband/core/mad.c 		if (atomic_read(&mad_agent_priv->qp_info->snoop_count))
atomic_read       549 drivers/infiniband/core/nldev.c 			      atomic_read(&cq->usecnt), RDMA_NLDEV_ATTR_PAD))
atomic_read       629 drivers/infiniband/core/nldev.c 			      atomic_read(&pd->usecnt), RDMA_NLDEV_ATTR_PAD))
atomic_read        99 drivers/infiniband/core/rdma_core.c 		WARN_ON(atomic_read(&uobj->usecnt) <= 0);
atomic_read       102 drivers/infiniband/core/rdma_core.c 		WARN_ON(atomic_read(&uobj->usecnt) != -1);
atomic_read       210 drivers/infiniband/core/security.c 			if (atomic_read(&pp->sec->error_list_count))
atomic_read       476 drivers/infiniband/core/security.c 	sec->error_comps_pending = atomic_read(&sec->error_list_count);
atomic_read        98 drivers/infiniband/core/uverbs_std_types_counters.c 	if (!atomic_read(&counters->usecnt))
atomic_read       340 drivers/infiniband/core/verbs.c 	WARN_ON(atomic_read(&pd->usecnt));
atomic_read      1031 drivers/infiniband/core/verbs.c 	if (atomic_read(&srq->usecnt))
atomic_read      1839 drivers/infiniband/core/verbs.c 	if (atomic_read(&real_qp->usecnt) == 0)
atomic_read      1869 drivers/infiniband/core/verbs.c 	if (atomic_read(&qp->usecnt))
atomic_read      1962 drivers/infiniband/core/verbs.c 	if (atomic_read(&cq->usecnt))
atomic_read      2244 drivers/infiniband/core/verbs.c 	if (atomic_read(&xrcd->usecnt))
atomic_read      2308 drivers/infiniband/core/verbs.c 	if (atomic_read(&wq->usecnt))
atomic_read      2389 drivers/infiniband/core/verbs.c 	if (atomic_read(&rwq_ind_table->usecnt))
atomic_read       126 drivers/infiniband/hw/bnxt_re/hw_counters.c 	stats->value[BNXT_RE_ACTIVE_QP] = atomic_read(&rdev->qp_count);
atomic_read       127 drivers/infiniband/hw/bnxt_re/hw_counters.c 	stats->value[BNXT_RE_ACTIVE_SRQ] = atomic_read(&rdev->srq_count);
atomic_read       128 drivers/infiniband/hw/bnxt_re/hw_counters.c 	stats->value[BNXT_RE_ACTIVE_CQ] = atomic_read(&rdev->cq_count);
atomic_read       129 drivers/infiniband/hw/bnxt_re/hw_counters.c 	stats->value[BNXT_RE_ACTIVE_MR] = atomic_read(&rdev->mr_count);
atomic_read       130 drivers/infiniband/hw/bnxt_re/hw_counters.c 	stats->value[BNXT_RE_ACTIVE_MW] = atomic_read(&rdev->mw_count);
atomic_read      1648 drivers/infiniband/hw/bnxt_re/main.c 		if (atomic_read(&rdev->sched_count) > 0)
atomic_read       163 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		if (atomic_read(&cq->arm_state) && nq->cqn_handler) {
atomic_read       100 drivers/infiniband/hw/cxgb3/iwch_provider.c 	wait_event(chp->wait, !atomic_read(&chp->refcnt));
atomic_read       682 drivers/infiniband/hw/cxgb3/iwch_provider.c 	wait_event(qhp->wait, !atomic_read(&qhp->refcnt));
atomic_read       980 drivers/infiniband/hw/cxgb4/cq.c 	wait_event(chp->wait, !atomic_read(&chp->refcnt));
atomic_read       132 drivers/infiniband/hw/cxgb4/device.c 	idx = atomic_read(&dev->rdev.wr_log_idx) &
atomic_read      1279 drivers/infiniband/hw/hfi1/driver.c 	if (atomic_read(&ppd->led_override_timer_active)) {
atomic_read       325 drivers/infiniband/hw/hfi1/file_ops.c 	if (atomic_read(&pq->n_reqs) == pq->n_max_reqs) {
atomic_read       217 drivers/infiniband/hw/hfi1/iowait.h 	wait_event(wait->wait_dma, !atomic_read(&wait->sdma_busy));
atomic_read       228 drivers/infiniband/hw/hfi1/iowait.h 	return atomic_read(&wait->sdma_busy);
atomic_read       271 drivers/infiniband/hw/hfi1/iowait.h 			   !atomic_read(&wait->pio_busy),
atomic_read       283 drivers/infiniband/hw/hfi1/iowait.h 	return atomic_read(&wait->pio_busy);
atomic_read       878 drivers/infiniband/hw/hfi1/mad.c 	is_beaconing_active = !!atomic_read(&ppd->led_override_timer_active);
atomic_read      4060 drivers/infiniband/hw/hfi1/mad.c 	is_beaconing_active = !!atomic_read(&ppd->led_override_timer_active);
atomic_read       671 drivers/infiniband/hw/hfi1/qp.c 		   atomic_read(&qp->refcount),
atomic_read       668 drivers/infiniband/hw/hfi1/rc.c 				if (atomic_read(&priv->n_tid_requests) >=
atomic_read      3369 drivers/infiniband/hw/hfi1/sdma.c 				       atomic_read(&dd->sdma_unfreeze_count) <=
atomic_read      3372 drivers/infiniband/hw/hfi1/sdma.c 	if (ret || atomic_read(&dd->sdma_unfreeze_count) < 0)
atomic_read      3388 drivers/infiniband/hw/hfi1/sdma.c 				atomic_read(&dd->sdma_unfreeze_count) <= 0);
atomic_read      5014 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (((atomic_read(&priv->n_tid_requests) < HFI1_TID_RDMA_WRITE_CNT) &&
atomic_read      5015 drivers/infiniband/hw/hfi1/tid_rdma.c 	     atomic_read(&priv->n_requests) &&
atomic_read      1500 drivers/infiniband/hw/hfi1/trace_tid.h 		__entry->n_requests = atomic_read(&priv->n_requests);
atomic_read      1501 drivers/infiniband/hw/hfi1/trace_tid.h 		__entry->n_tid_requests = atomic_read(&priv->n_tid_requests);
atomic_read       297 drivers/infiniband/hw/hfi1/user_sdma.c 			!atomic_read(&pq->n_reqs));
atomic_read       984 drivers/infiniband/hw/hfi1/user_sdma.c 				atomic_read(&pq->n_locked), npages)) {
atomic_read      1514 drivers/infiniband/hw/hfi1/user_sdma.c 	if (atomic_read(&node->refcount))
atomic_read      1541 drivers/infiniband/hw/hfi1/user_sdma.c 	if (!atomic_read(&node->refcount))
atomic_read      1948 drivers/infiniband/hw/i40iw/i40iw_cm.c 		if (atomic_read(&listener->pend_accepts_cnt) > 0)
atomic_read      1954 drivers/infiniband/hw/i40iw/i40iw_cm.c 				    atomic_read(&listener->pend_accepts_cnt));
atomic_read      2598 drivers/infiniband/hw/i40iw/i40iw_cm.c 		if (atomic_read(&cm_node->listener->pend_accepts_cnt) >
atomic_read       362 drivers/infiniband/hw/i40iw/i40iw_hw.c 			if (atomic_read(&iwqp->close_timer_started))
atomic_read      1763 drivers/infiniband/hw/i40iw/i40iw_main.c 	if (atomic_read(&iwdev->params_busy))
atomic_read      1977 drivers/infiniband/hw/i40iw/i40iw_main.c 	    (atomic_read(&iwdev->vchnl_msgs) == 0))
atomic_read      1984 drivers/infiniband/hw/i40iw/i40iw_main.c 				(atomic_read(&iwdev->vchnl_msgs) == 0),
atomic_read       449 drivers/infiniband/hw/i40iw/i40iw_utils.c 			   !atomic_read(&cqp_request->refcount),
atomic_read      1170 drivers/infiniband/hw/i40iw/i40iw_utils.c 					 (atomic_read(&iwdev->vchnl_msgs) == 1),
atomic_read      1017 drivers/infiniband/hw/mlx4/mcg.c 			atomic_read(&group->refcount),
atomic_read      1104 drivers/infiniband/hw/mlx4/mcg.c 		if (atomic_read(&group->refcount))
atomic_read      1106 drivers/infiniband/hw/mlx4/mcg.c 					atomic_read(&group->refcount), group);
atomic_read      1234 drivers/infiniband/hw/mlx4/mcg.c 		if (atomic_read(&group->refcount)) {
atomic_read      3210 drivers/infiniband/hw/mlx5/main.c 	    atomic_read(&handler->ibcounters->usecnt) == 1)
atomic_read      3641 drivers/infiniband/hw/mlx5/main.c 		    atomic_read(&handler->ibcounters->usecnt) == 1)
atomic_read      4408 drivers/infiniband/hw/mlx5/main.c 	return sprintf(buf, "%d\n", atomic_read(&dev->mdev->priv.reg_pages));
atomic_read      1590 drivers/infiniband/hw/mlx5/mr.c 		if (atomic_read(&mr->num_pending_prefetch))
atomic_read      1592 drivers/infiniband/hw/mlx5/mr.c 		WARN_ON(atomic_read(&mr->num_pending_prefetch));
atomic_read       606 drivers/infiniband/hw/mlx5/odp.c 	wait_event(imr->q_leaf_free, !atomic_read(&imr->num_leaf_free));
atomic_read       170 drivers/infiniband/hw/mthca/mthca_av.c 	} else if (!atomic_read(&pd->sqp_count) &&
atomic_read       774 drivers/infiniband/hw/qib/qib_driver.c 		if (atomic_read(&ppd->led_override_timer_active)) {
atomic_read       359 drivers/infiniband/hw/qib/qib_qp.c 	wait_event(priv->wait_dma, !atomic_read(&priv->s_dma_busy));
atomic_read       440 drivers/infiniband/hw/qib/qib_qp.c 		   atomic_read(&priv->s_dma_busy),
atomic_read       247 drivers/infiniband/hw/qib/qib_rc.c 		if (atomic_read(&priv->s_dma_busy)) {
atomic_read        66 drivers/infiniband/hw/qib/qib_uc.c 		if (atomic_read(&priv->s_dma_busy)) {
atomic_read       253 drivers/infiniband/hw/qib/qib_ud.c 		if (atomic_read(&priv->s_dma_busy)) {
atomic_read       293 drivers/infiniband/hw/qib/qib_ud.c 			if (atomic_read(&priv->s_dma_busy)) {
atomic_read        71 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	const unsigned int idx = atomic_read(var);
atomic_read        80 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	__u32 idx = atomic_read(var) + 1;	/* Increment. */
atomic_read        89 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	const __u32 tail = atomic_read(&r->prod_tail);
atomic_read        90 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	const __u32 head = atomic_read(&r->cons_head);
atomic_read       103 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	const __u32 tail = atomic_read(&r->prod_tail);
atomic_read       104 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	const __u32 head = atomic_read(&r->cons_head);
atomic_read       413 drivers/infiniband/sw/rdmavt/mcast.c 	wait_event(mcast->wait, atomic_read(&mcast->refcount) <= 1);
atomic_read       418 drivers/infiniband/sw/rdmavt/mcast.c 		wait_event(mcast->wait, !atomic_read(&mcast->refcount));
atomic_read       958 drivers/infiniband/sw/rdmavt/mr.c 	if (unlikely(atomic_read(&mr->lkey_invalid) ||
atomic_read      1066 drivers/infiniband/sw/rdmavt/mr.c 	if (unlikely(atomic_read(&mr->lkey_invalid) ||
atomic_read       304 drivers/infiniband/sw/rdmavt/qp.c 	return atomic_read(&wss->total_count) >= wss->threshold;
atomic_read      1724 drivers/infiniband/sw/rdmavt/qp.c 	wait_event(qp->wait, !atomic_read(&qp->refcount));
atomic_read      1962 drivers/infiniband/sw/rdmavt/qp.c 		reserved_used = atomic_read(&qp->s_reserved_used);
atomic_read      1977 drivers/infiniband/sw/rdmavt/qp.c 	reserved_used = atomic_read(&qp->s_reserved_used);
atomic_read      2051 drivers/infiniband/sw/rdmavt/qp.c 			    atomic_read(&qp->local_ops_pending)) {
atomic_read       260 drivers/infiniband/sw/rxe/rxe_pool.c 	if (atomic_read(&pool->num_elem) > 0)
atomic_read       673 drivers/infiniband/sw/rxe/rxe_req.c 	if (unlikely(atomic_read(&qp->skb_out) >
atomic_read       971 drivers/infiniband/sw/siw/siw_cm.c 	if (atomic_read(&new_s->sk->sk_rmem_alloc)) {
atomic_read       214 drivers/infiniband/sw/siw/siw_main.c 		usage = atomic_read(&per_cpu(siw_use_cnt, cpu));
atomic_read       123 drivers/infiniband/sw/siw/siw_verbs.c 		atomic_read(&sdev->num_ctx));
atomic_read       130 drivers/infiniband/sw/siw/siw_verbs.c 		atomic_read(&sdev->num_ctx));
atomic_read       273 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "now %d PD's(s)\n", atomic_read(&sdev->num_pd));
atomic_read      1613 drivers/infiniband/ulp/ipoib/ipoib_main.c 	wait_flushed = atomic_read(&priv->ntbl.entries);
atomic_read      2763 drivers/infiniband/ulp/srpt/ib_srpt.c 			 atomic_read(&ch->sq_wr_avail));
atomic_read      2610 drivers/input/mouse/cyapa_gen5.c 	if (atomic_read(&pip->cmd_issued)) {
atomic_read      1078 drivers/input/touchscreen/cyttsp4_core.c 	if (atomic_read(&cd->ignore_irq)) {
atomic_read       269 drivers/iommu/amd_iommu_v2.c 	wait_event(pasid_state->wq, !atomic_read(&pasid_state->count));
atomic_read       856 drivers/iommu/amd_iommu_v2.c 	wait_event(dev_state->wq, !atomic_read(&dev_state->count));
atomic_read      1031 drivers/iommu/arm-smmu-v3.c 	if (atomic_read(&cmdq->lock) == 1)
atomic_read      1946 drivers/iommu/arm-smmu-v3.c 	if (!atomic_read(&smmu_domain->nr_ats_masters))
atomic_read       464 drivers/iommu/intel-svm.c 	else if (atomic_read(&svm->mm->mm_users) > 0)
atomic_read       581 drivers/iommu/iova.c 	if (!atomic_read(&iovad->fq_timer_on) &&
atomic_read       579 drivers/irqchip/irq-renesas-intc-irqpin.c 	if (atomic_read(&p->wakeup_path))
atomic_read       233 drivers/irqchip/irq-renesas-irqc.c 	if (atomic_read(&p->wakeup_path))
atomic_read        79 drivers/leds/trigger/ledtrig-cpu.c 		active_cpus = atomic_read(&num_active_cpus);
atomic_read       258 drivers/leds/trigger/ledtrig-netdev.c 		       jiffies_to_msecs(atomic_read(&trigger_data->interval)));
atomic_read       379 drivers/leds/trigger/ledtrig-netdev.c 				atomic_read(&trigger_data->interval));
atomic_read       389 drivers/leds/trigger/ledtrig-netdev.c 			(atomic_read(&trigger_data->interval)*2));
atomic_read       388 drivers/lightnvm/pblk-core.c 		if (!atomic_read(&pblk->inflight_io))
atomic_read      1138 drivers/lightnvm/pblk-core.c 	int blk_to_erase = atomic_read(&line->blk_in_line);
atomic_read      1160 drivers/lightnvm/pblk-core.c 	int blk_in_line = atomic_read(&line->blk_in_line);
atomic_read      1561 drivers/lightnvm/pblk-core.c 	left_seblks = atomic_read(&new->left_seblks);
atomic_read      1564 drivers/lightnvm/pblk-core.c 		if (atomic_read(&new->left_eblks)) {
atomic_read       380 drivers/lightnvm/pblk-gc.c 		if (!atomic_read(&line->sec_to_update))
atomic_read       397 drivers/lightnvm/pblk-gc.c 	unsigned int werr_lines = atomic_read(&rl->werr_lines);
atomic_read       456 drivers/lightnvm/pblk-gc.c 	if (!run_gc || (atomic_read(&gc->read_inflight_gc) >= PBLK_GC_L_QD))
atomic_read       549 drivers/lightnvm/pblk-gc.c 		atomic_read(&gc->pipeline_gc));
atomic_read       553 drivers/lightnvm/pblk-gc.c 		if (!atomic_read(&gc->pipeline_gc))
atomic_read       830 drivers/lightnvm/pblk-rb.c 			atomic_read(&rb->inflight_flush_point),
atomic_read       848 drivers/lightnvm/pblk-rb.c 			atomic_read(&rb->inflight_flush_point),
atomic_read        31 drivers/lightnvm/pblk-rl.c 	rb_space = atomic_read(&rl->rb_space);
atomic_read        38 drivers/lightnvm/pblk-rl.c 	int rb_user_cnt = atomic_read(&rl->rb_user_cnt);
atomic_read        39 drivers/lightnvm/pblk-rl.c 	int rb_space = atomic_read(&rl->rb_space);
atomic_read        52 drivers/lightnvm/pblk-rl.c 	int rb_space = atomic_read(&rl->rb_space);
atomic_read        60 drivers/lightnvm/pblk-rl.c 	int rb_gc_cnt = atomic_read(&rl->rb_gc_cnt);
atomic_read       100 drivers/lightnvm/pblk-rl.c 	return atomic_read(&rl->free_blocks);
atomic_read       105 drivers/lightnvm/pblk-rl.c 	return atomic_read(&rl->free_user_blocks);
atomic_read       113 drivers/lightnvm/pblk-rl.c 	int werr_gc_needed = atomic_read(&rl->werr_lines);
atomic_read       161 drivers/lightnvm/pblk-rl.c 	int blk_in_line = atomic_read(&line->blk_in_line);
atomic_read       173 drivers/lightnvm/pblk-rl.c 	int blk_in_line = atomic_read(&line->blk_in_line);
atomic_read       182 drivers/lightnvm/pblk-rl.c 		free_blocks = atomic_read(&rl->free_user_blocks);
atomic_read        60 drivers/lightnvm/pblk-sysfs.c 	rb_user_cnt = atomic_read(&pblk->rl.rb_user_cnt);
atomic_read        62 drivers/lightnvm/pblk-sysfs.c 	rb_gc_cnt = atomic_read(&pblk->rl.rb_gc_cnt);
atomic_read       298 drivers/lightnvm/pblk-sysfs.c 			atomic_read(&pblk->gc.read_inflight_gc));
atomic_read       304 drivers/lightnvm/pblk-sysfs.c 			atomic_read(&pblk->inflight_io));
atomic_read       338 drivers/lightnvm/pblk-write.c 	if (likely(!e_line || !atomic_read(&e_line->left_eblks)))
atomic_read       627 drivers/macintosh/adb.c 		if (atomic_read(&state->n_pending) == 0) {
atomic_read       701 drivers/macintosh/adb.c 		if (atomic_read(&state->n_pending) == 0
atomic_read       737 drivers/macintosh/adb.c 		else if (atomic_read(&state->n_pending) == 0)
atomic_read       980 drivers/mailbox/bcm-flexrm-mailbox.c 			   (u32)atomic_read(&ring->msg_send_count),
atomic_read       981 drivers/mailbox/bcm-flexrm-mailbox.c 			   (u32)atomic_read(&ring->msg_cmpl_count));
atomic_read        97 drivers/md/bcache/alloc.c 		r = atomic_read(&c->rescale);
atomic_read       111 drivers/md/bcache/alloc.c 			    !atomic_read(&b->pin)) {
atomic_read       137 drivers/md/bcache/alloc.c 		!atomic_read(&b->pin) &&
atomic_read       364 drivers/md/bcache/alloc.c 		allocator_wait(ca, !atomic_read(&ca->set->prio_blocked));
atomic_read       448 drivers/md/bcache/alloc.c 	BUG_ON(atomic_read(&b->pin) != 1);
atomic_read      1675 drivers/md/bcache/btree.c 		if (atomic_read(&b->c->search_inflight) &&
atomic_read      1752 drivers/md/bcache/btree.c 			if (!atomic_read(&b->pin)) {
atomic_read      1815 drivers/md/bcache/btree.c 			if (atomic_read(&b->pin))
atomic_read      1881 drivers/md/bcache/btree.c 	if (atomic_read(&c->sectors_to_gc) < 0)
atomic_read        90 drivers/md/bcache/closure.c 	if (atomic_read(&cl->remaining) & CLOSURE_WAITING)
atomic_read       176 drivers/md/bcache/closure.c 		int r = atomic_read(&cl->remaining);
atomic_read       183 drivers/md/bcache/closure.h 	if ((atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK) != 1)
atomic_read       200 drivers/md/bcache/extents.c 		  buf, PTR_BUCKET_NR(b->c, k, i), atomic_read(&g->pin),
atomic_read       532 drivers/md/bcache/extents.c 		  buf, PTR_BUCKET_NR(b->c, k, ptr), atomic_read(&g->pin),
atomic_read       109 drivers/md/bcache/io.c 				errors = atomic_read(&ca->io_errors);
atomic_read       150 drivers/md/bcache/io.c 		int congested = atomic_read(&c->congested);
atomic_read       369 drivers/md/bcache/journal.c 		BUG_ON(i->pin && atomic_read(i->pin) != 1);
atomic_read       444 drivers/md/bcache/journal.c 	ref_nr = atomic_read(fifo_front_p);
atomic_read       481 drivers/md/bcache/journal.c 		ref_nr = atomic_read(fifo_front_p);
atomic_read       608 drivers/md/bcache/journal.c 	switch (atomic_read(&ja->discard_in_flight)) {
atomic_read       649 drivers/md/bcache/journal.c 	while (!atomic_read(&fifo_front(&c->journal.pin)))
atomic_read       219 drivers/md/bcache/movinggc.c 			    atomic_read(&b->pin))
atomic_read        73 drivers/md/bcache/request.c 	while (atomic_read(&s->cl.remaining) & CLOSURE_WAITING)
atomic_read       348 drivers/md/bcache/request.c 	i += atomic_read(&c->congested);
atomic_read       620 drivers/md/bcache/request.c 			if (dc && atomic_read(&dc->has_dirty))
atomic_read      1192 drivers/md/bcache/request.c 		if (atomic_read(&d->c->idle_counter))
atomic_read      1200 drivers/md/bcache/request.c 		if (unlikely(atomic_read(&d->c->at_max_writeback_rate) == 1)) {
atomic_read       182 drivers/md/bcache/stats.c 	if (!atomic_read(&acc->closing))
atomic_read      1270 drivers/md/bcache/super.c 	if (atomic_read(&dc->running))
atomic_read      1678 drivers/md/bcache/super.c 	} else if (atomic_read(&dc->has_dirty)) {
atomic_read       197 drivers/md/bcache/sysfs.c 	sysfs_printf(io_errors,		"%i", atomic_read(&dc->io_errors));
atomic_read       252 drivers/md/bcache/sysfs.c 	sysfs_print(running,		atomic_read(&dc->running));
atomic_read      1009 drivers/md/bcache/sysfs.c 		    atomic_read(&ca->io_errors) >> IO_ERROR_SHIFT);
atomic_read       142 drivers/md/bcache/writeback.c 	    atomic_read(&c->attached_dev_nr) * 6)
atomic_read       145 drivers/md/bcache/writeback.c 	if (atomic_read(&c->at_max_writeback_rate) != 1)
atomic_read       161 drivers/md/bcache/writeback.c 	if ((atomic_read(&c->idle_counter) <
atomic_read       162 drivers/md/bcache/writeback.c 	     atomic_read(&c->attached_dev_nr) * 6) ||
atomic_read       163 drivers/md/bcache/writeback.c 	    !atomic_read(&c->at_max_writeback_rate))
atomic_read       196 drivers/md/bcache/writeback.c 	if (atomic_read(&dc->has_dirty) && dc->writeback_percent) {
atomic_read       330 drivers/md/bcache/writeback.c 	if (atomic_read(&dc->writeback_sequence_next) != io->sequence) {
atomic_read       334 drivers/md/bcache/writeback.c 		if (atomic_read(&dc->writeback_sequence_next) == io->sequence) {
atomic_read       674 drivers/md/bcache/writeback.c 		    (!atomic_read(&dc->has_dirty) || !dc->writeback_running)) {
atomic_read       775 drivers/md/bcache/writeback.c 	if (atomic_read(&b->c->search_inflight) &&
atomic_read        31 drivers/md/bcache/writeback.h 		ret += atomic_read(d->stripe_sectors_dirty + i);
atomic_read        50 drivers/md/bcache/writeback.h 		if (atomic_read(dc->disk.stripe_sectors_dirty + stripe))
atomic_read        98 drivers/md/bcache/writeback.h 	if (!atomic_read(&dc->has_dirty) &&
atomic_read        38 drivers/md/dm-bio-record.h 	bd->__bi_remaining = atomic_read(&bio->__bi_remaining);
atomic_read       152 drivers/md/dm-cache-background-tracker.c 	return atomic_read(&b->pending_writebacks);
atomic_read       158 drivers/md/dm-cache-background-tracker.c 	return atomic_read(&b->pending_demotes);
atomic_read       164 drivers/md/dm-cache-background-tracker.c 	return atomic_read(&b->pending_promotes) +
atomic_read       165 drivers/md/dm-cache-background-tracker.c 		atomic_read(&b->pending_writebacks) +
atomic_read       166 drivers/md/dm-cache-background-tracker.c 		atomic_read(&b->pending_demotes) >= b->max_work;
atomic_read      1054 drivers/md/dm-cache-target.c 	stats.read_hits = atomic_read(&cache->stats.read_hit);
atomic_read      1055 drivers/md/dm-cache-target.c 	stats.read_misses = atomic_read(&cache->stats.read_miss);
atomic_read      1056 drivers/md/dm-cache-target.c 	stats.write_hits = atomic_read(&cache->stats.write_hit);
atomic_read      1057 drivers/md/dm-cache-target.c 	stats.write_misses = atomic_read(&cache->stats.write_miss);
atomic_read      1672 drivers/md/dm-cache-target.c 	sector_t current_volume = (atomic_read(&cache->nr_io_migrations) + 1) *
atomic_read      2868 drivers/md/dm-cache-target.c 	BUG_ON(atomic_read(&cache->nr_io_migrations));
atomic_read      3198 drivers/md/dm-cache-target.c 		       (unsigned) atomic_read(&cache->stats.read_hit),
atomic_read      3199 drivers/md/dm-cache-target.c 		       (unsigned) atomic_read(&cache->stats.read_miss),
atomic_read      3200 drivers/md/dm-cache-target.c 		       (unsigned) atomic_read(&cache->stats.write_hit),
atomic_read      3201 drivers/md/dm-cache-target.c 		       (unsigned) atomic_read(&cache->stats.write_miss),
atomic_read      3202 drivers/md/dm-cache-target.c 		       (unsigned) atomic_read(&cache->stats.demotion),
atomic_read      3203 drivers/md/dm-cache-target.c 		       (unsigned) atomic_read(&cache->stats.promotion),
atomic_read      3204 drivers/md/dm-cache-target.c 		       (unsigned long) atomic_read(&cache->nr_dirty));
atomic_read       790 drivers/md/dm-clone-target.c 	    !atomic_read(&clone->ios_in_flight))
atomic_read      1097 drivers/md/dm-clone-target.c 	       !atomic_read(&clone->ios_in_flight) &&
atomic_read      1100 drivers/md/dm-clone-target.c 		current_volume = atomic_read(&clone->hydrations_in_flight);
atomic_read      1479 drivers/md/dm-clone-target.c 		       atomic_read(&clone->hydrations_in_flight));
atomic_read      2024 drivers/md/dm-clone-target.c 	wait_event(clone->hydration_stopped, !atomic_read(&clone->hydrations_in_flight));
atomic_read       248 drivers/md/dm-delay.c 	if (!c->delay || !atomic_read(&dc->may_delay))
atomic_read      1204 drivers/md/dm-era-target.c 	if (!atomic_read(&era->suspended))
atomic_read       777 drivers/md/dm-integrity.c 	BUG_ON(!atomic_read(&comp->in_flight));
atomic_read      1255 drivers/md/dm-ioctl.c 	priv->global_event_nr = atomic_read(&dm_global_event_nr);
atomic_read      1942 drivers/md/dm-ioctl.c 	priv->global_event_nr = atomic_read(&dm_global_event_nr);
atomic_read      1960 drivers/md/dm-ioctl.c 	if ((int)(atomic_read(&dm_global_event_nr) - priv->global_event_nr) > 0)
atomic_read       973 drivers/md/dm-kcopyd.c 	wait_event(kc->destroyq, !atomic_read(&kc->nr_jobs));
atomic_read       332 drivers/md/dm-log-userspace-base.c 		if (atomic_read(&lc->sched_flush))
atomic_read       371 drivers/md/dm-log-userspace-base.c 	if (lc->integrated_flush && atomic_read(&lc->sched_flush))
atomic_read       605 drivers/md/dm-log-userspace-base.c 	if (mark_list_is_empty && !atomic_read(&lc->sched_flush)) {
atomic_read       647 drivers/md/dm-log-writes.c 	wait_event(lc->wait, !atomic_read(&lc->io_blocks) &&
atomic_read       648 drivers/md/dm-log-writes.c 		   !atomic_read(&lc->pending_blocks));
atomic_read       291 drivers/md/dm-mpath.c 	if (atomic_read(&m->pg_init_in_progress) || test_bit(MPATHF_PG_INIT_DISABLED, &m->flags))
atomic_read       312 drivers/md/dm-mpath.c 	return atomic_read(&m->pg_init_in_progress);
atomic_read       375 drivers/md/dm-mpath.c 	if (!atomic_read(&m->nr_valid_paths)) {
atomic_read      1135 drivers/md/dm-mpath.c 		unsigned nr_valid_paths = atomic_read(&m->nr_valid_paths);
atomic_read      1182 drivers/md/dm-mpath.c 		if (!atomic_read(&m->pg_init_in_progress))
atomic_read      1196 drivers/md/dm-mpath.c 		if (atomic_read(&m->pg_init_in_progress))
atomic_read      1242 drivers/md/dm-mpath.c 		       pgpath->path.dev->name, atomic_read(&m->nr_valid_paths));
atomic_read      1405 drivers/md/dm-mpath.c 	if (atomic_read(&m->pg_init_count) <= m->pg_init_retries &&
atomic_read      1550 drivers/md/dm-mpath.c 		if (atomic_read(&m->nr_valid_paths) == 0 &&
atomic_read      1584 drivers/md/dm-mpath.c 	if (atomic_read(&m->nr_valid_paths) == 0 &&
atomic_read      1681 drivers/md/dm-mpath.c 		       atomic_read(&m->pg_init_count));
atomic_read      1940 drivers/md/dm-mpath.c 	if (atomic_read(&m->pg_init_in_progress))
atomic_read      1944 drivers/md/dm-mpath.c 	if (!atomic_read(&m->nr_valid_paths) && test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags))
atomic_read       100 drivers/md/dm-queue-length.c 			DMEMIT("%d ", atomic_read(&pi->qlen));
atomic_read       200 drivers/md/dm-queue-length.c 		    (atomic_read(&pi->qlen) < atomic_read(&best->qlen)))
atomic_read       203 drivers/md/dm-queue-length.c 		if (!atomic_read(&best->qlen))
atomic_read       175 drivers/md/dm-raid1.c 	return &ms->mirror[atomic_read(&ms->default_mirror)];
atomic_read       191 drivers/md/dm-raid1.c 		if (!atomic_read(&m->error_count))
atomic_read       420 drivers/md/dm-raid1.c 		if (likely(!atomic_read(&m->error_count)))
atomic_read       434 drivers/md/dm-raid1.c 	return !atomic_read(&default_mirror->error_count);
atomic_read       480 drivers/md/dm-raid1.c 	if (atomic_read(&ms->suspend)) {
atomic_read       574 drivers/md/dm-raid1.c 		else if (m && atomic_read(&m->error_count))
atomic_read      1383 drivers/md/dm-raid1.c 	if (!atomic_read(&(m->error_count)))
atomic_read       246 drivers/md/dm-region-hash.c 			BUG_ON(atomic_read(&reg->pending));
atomic_read       612 drivers/md/dm-region-hash.c 	if (atomic_read(&reg->pending))
atomic_read       680 drivers/md/dm-region-hash.c 	return atomic_read(&rh->recovery_in_flight);
atomic_read        95 drivers/md/dm-service-time.c 			DMEMIT("%d %u ", atomic_read(&pi->in_flight_size),
atomic_read       217 drivers/md/dm-service-time.c 	sz1 = atomic_read(&pi1->in_flight_size);
atomic_read       218 drivers/md/dm-service-time.c 	sz2 = atomic_read(&pi2->in_flight_size);
atomic_read      1488 drivers/md/dm-snap.c 	while (atomic_read(&s->pending_exceptions_count))
atomic_read       187 drivers/md/dm-stats.c 	return atomic_read(&shared->in_flight[READ]) +
atomic_read       188 drivers/md/dm-stats.c 	       atomic_read(&shared->in_flight[WRITE]);
atomic_read       224 drivers/md/dm-stats.c 				       atomic_read(&shared->in_flight[READ]),
atomic_read       225 drivers/md/dm-stats.c 				       atomic_read(&shared->in_flight[WRITE]));
atomic_read       501 drivers/md/dm-stats.c 	in_flight_read = (unsigned)atomic_read(&shared->in_flight[READ]);
atomic_read       502 drivers/md/dm-stats.c 	in_flight_write = (unsigned)atomic_read(&shared->in_flight[WRITE]);
atomic_read       410 drivers/md/dm-stripe.c 			DMEMIT("%c", atomic_read(&(sc->stripe[i].error_count)) ?
atomic_read       453 drivers/md/dm-stripe.c 			if (atomic_read(&(sc->stripe[i].error_count)) <
atomic_read       440 drivers/md/dm-writecache.c 	BUG_ON(atomic_read(&endio->count) <= 0);
atomic_read       448 drivers/md/dm-writecache.c 		   !atomic_read(&wc->bio_in_progress[direction]));
atomic_read       217 drivers/md/dm-zoned-metadata.c 	return atomic_read(&zmd->unmap_nr_rnd);
atomic_read       275 drivers/md/dm-zoned-metadata.c 	if (zmd->max_nr_mblks && atomic_read(&zmd->nr_mblks) > zmd->max_nr_mblks) {
atomic_read       466 drivers/md/dm-zoned-metadata.c 	       atomic_read(&zmd->nr_mblks) > zmd->min_nr_mblks &&
atomic_read       487 drivers/md/dm-zoned-metadata.c 	return atomic_read(&zmd->nr_mblks);
atomic_read      1457 drivers/md/dm-zoned-metadata.c 		} else if (atomic_read(&zmd->nr_reserved_seq_zones) < zmd->nr_reserved_seq) {
atomic_read      1855 drivers/md/dm-zoned-metadata.c 	} else if (atomic_read(&zmd->nr_reserved_seq_zones) <
atomic_read      2486 drivers/md/dm-zoned-metadata.c 		     zmd->nr_rnd, atomic_read(&zmd->unmap_nr_rnd));
atomic_read      2488 drivers/md/dm-zoned-metadata.c 		     zmd->nr_seq, atomic_read(&zmd->unmap_nr_seq));
atomic_read       218 drivers/md/dm-zoned.h 	return atomic_read(&zone->refcount);
atomic_read       374 drivers/md/dm.c 	return atomic_read(&md->open_count);
atomic_read      2261 drivers/md/dm.c 	BUG_ON(!atomic_read(&md->holders));
atomic_read      2406 drivers/md/dm.c 		while (atomic_read(&md->holders))
atomic_read      2408 drivers/md/dm.c 	else if (atomic_read(&md->holders))
atomic_read      2410 drivers/md/dm.c 		       dm_device_name(md), atomic_read(&md->holders));
atomic_read      2909 drivers/md/dm.c 	return atomic_read(&md->event_nr);
atomic_read      2915 drivers/md/dm.c 			(event_nr != atomic_read(&md->event_nr)));
atomic_read       311 drivers/md/md-bitmap.c 				   atomic_read(&bitmap->pending_writes)==0);
atomic_read       408 drivers/md/md-bitmap.c 		   atomic_read(&bitmap->pending_writes)==0);
atomic_read       433 drivers/md/md-bitmap.c 			   atomic_read(&bitmap->pending_writes)==0);
atomic_read      1404 drivers/md/md-bitmap.c 		bw = atomic_read(&bitmap->behind_writes);
atomic_read      1469 drivers/md/md-bitmap.c 			 atomic_read(&bitmap->behind_writes),
atomic_read      1628 drivers/md/md-bitmap.c 		   atomic_read(&bitmap->mddev->recovery_active) == 0);
atomic_read      1752 drivers/md/md-bitmap.c 		   atomic_read(&bitmap->pending_writes) == 0);
atomic_read      1776 drivers/md/md-bitmap.c 	if (bitmap && atomic_read(&bitmap->behind_writes) > 0) {
atomic_read      1781 drivers/md/md-bitmap.c 			   atomic_read(&bitmap->behind_writes) == 0);
atomic_read        88 drivers/md/md-faulty.c 	    atomic_read(&conf->counters[mode]) <= 0)
atomic_read       171 drivers/md/md-faulty.c 		if (atomic_read(&conf->counters[WriteAll])) {
atomic_read       226 drivers/md/md-faulty.c 	if ((n=atomic_read(&conf->counters[WriteTransient])) != 0)
atomic_read       230 drivers/md/md-faulty.c 	if ((n=atomic_read(&conf->counters[ReadTransient])) != 0)
atomic_read       234 drivers/md/md-faulty.c 	if ((n=atomic_read(&conf->counters[WritePersistent])) != 0)
atomic_read       238 drivers/md/md-faulty.c 	if ((n=atomic_read(&conf->counters[ReadPersistent])) != 0)
atomic_read       243 drivers/md/md-faulty.c 	if ((n=atomic_read(&conf->counters[ReadFixable])) != 0)
atomic_read       247 drivers/md/md-faulty.c 	if ((n=atomic_read(&conf->counters[WriteAll])) != 0)
atomic_read       282 drivers/md/md-multipath.c 		    atomic_read(&rdev->nr_pending)) {
atomic_read       290 drivers/md/md-multipath.c 			if (atomic_read(&rdev->nr_pending)) {
atomic_read       431 drivers/md/md.c 	wait_event(mddev->sb_wait, atomic_read(&mddev->active_io) == 0);
atomic_read       910 drivers/md/md.c 	wait_event(mddev->sb_wait, atomic_read(&mddev->pending_writes)==0);
atomic_read      1923 drivers/md/md.c 	sb->cnt_corrected_read = cpu_to_le32(atomic_read(&rdev->corrected_errors));
atomic_read      3012 drivers/md/md.c 	return sprintf(page, "%d\n", atomic_read(&rdev->corrected_errors));
atomic_read      4398 drivers/md/md.c 		       atomic_read(&mddev->max_corr_read_errors));
atomic_read      4865 drivers/md/md.c 	resync = mddev->curr_mark_cnt - atomic_read(&mddev->recovery_active);
atomic_read      6109 drivers/md/md.c 	if ((mddev->pers && atomic_read(&mddev->openers) > !!bdev) ||
atomic_read      6170 drivers/md/md.c 	if ((mddev->pers && atomic_read(&mddev->openers) > !!bdev) ||
atomic_read      6848 drivers/md/md.c 		} else if (atomic_read(&inode->i_writecount) != 1) {
atomic_read      7368 drivers/md/md.c 		if (mddev->pers && atomic_read(&mddev->openers) > 1) {
atomic_read      7833 drivers/md/md.c 		resync -= atomic_read(&mddev->recovery_active);
atomic_read      7921 drivers/md/md.c 	recovery_active = atomic_read(&mddev->recovery_active);
atomic_read      8016 drivers/md/md.c 		seq->poll_event = atomic_read(&md_event_count);
atomic_read      8116 drivers/md/md.c 	seq->poll_event = atomic_read(&md_event_count);
atomic_read      8133 drivers/md/md.c 	if (seq->poll_event != atomic_read(&md_event_count))
atomic_read      8228 drivers/md/md.c 			      atomic_read(&disk->sync_io);
atomic_read      8628 drivers/md/md.c 				   atomic_read(&mddev->recovery_active) == 0);
atomic_read      8693 drivers/md/md.c 			mark_cnt[next] = io_sectors - atomic_read(&mddev->recovery_active);
atomic_read      8710 drivers/md/md.c 		recovery_done = io_sectors - atomic_read(&mddev->recovery_active);
atomic_read      8725 drivers/md/md.c 					   !atomic_read(&mddev->recovery_active));
atomic_read      8736 drivers/md/md.c 	wait_event(mddev->recovery_wait, !atomic_read(&mddev->recovery_active));
atomic_read      8838 drivers/md/md.c 		    atomic_read(&rdev->nr_pending)==0) {
atomic_read      8858 drivers/md/md.c 		    atomic_read(&rdev->nr_pending)==0)) {
atomic_read       518 drivers/md/raid1.c 		if (atomic_read(&r1_bio->behind_remaining) >= (atomic_read(&r1_bio->remaining)-1) &&
atomic_read       690 drivers/md/raid1.c 		pending = atomic_read(&rdev->nr_pending);
atomic_read       889 drivers/md/raid1.c 			    !atomic_read(&conf->nr_waiting[idx]),
atomic_read       913 drivers/md/raid1.c 			     !atomic_read(&conf->nr_pending[idx]) &&
atomic_read       914 drivers/md/raid1.c 			     atomic_read(&conf->barrier[idx]) < RESYNC_DEPTH) ||
atomic_read       935 drivers/md/raid1.c 	BUG_ON(atomic_read(&conf->barrier[idx]) <= 0);
atomic_read       973 drivers/md/raid1.c 	    !atomic_read(&conf->barrier[idx]))
atomic_read       994 drivers/md/raid1.c 			     !atomic_read(&conf->barrier[idx]),
atomic_read      1059 drivers/md/raid1.c 	ret = atomic_read(&conf->nr_sync_pending);
atomic_read      1061 drivers/md/raid1.c 		ret += atomic_read(&conf->nr_pending[idx]) -
atomic_read      1062 drivers/md/raid1.c 			atomic_read(&conf->nr_queued[idx]);
atomic_read      1290 drivers/md/raid1.c 			   atomic_read(&bitmap->behind_writes) == 0);
atomic_read      1490 drivers/md/raid1.c 			    (atomic_read(&bitmap->behind_writes)
atomic_read      1819 drivers/md/raid1.c 		    atomic_read(&rdev->nr_pending)) {
atomic_read      1835 drivers/md/raid1.c 			if (atomic_read(&rdev->nr_pending)) {
atomic_read      1850 drivers/md/raid1.c 			if (atomic_read(&repl->nr_pending)) {
atomic_read      2697 drivers/md/raid1.c 	if (atomic_read(&conf->nr_waiting[idx]))
atomic_read       800 drivers/md/raid10.c 		pending = atomic_read(&rdev->nr_pending);
atomic_read       964 drivers/md/raid10.c 			    !atomic_read(&conf->nr_pending) && conf->barrier < RESYNC_DEPTH,
atomic_read       996 drivers/md/raid10.c 				    (atomic_read(&conf->nr_pending) &&
atomic_read      1035 drivers/md/raid10.c 				atomic_read(&conf->nr_pending) == conf->nr_queued+extra,
atomic_read      1839 drivers/md/raid10.c 	    atomic_read(&rdev->nr_pending)) {
atomic_read      1857 drivers/md/raid10.c 		if (atomic_read(&rdev->nr_pending)) {
atomic_read      2272 drivers/md/raid10.c 	unsigned int read_errors = atomic_read(&rdev->read_errors);
atomic_read      2335 drivers/md/raid10.c 	int max_read_errors = atomic_read(&mddev->max_corr_read_errors);
atomic_read      2350 drivers/md/raid10.c 	if (atomic_read(&rdev->read_errors) > max_read_errors) {
atomic_read      2356 drivers/md/raid10.c 			  atomic_read(&rdev->read_errors), max_read_errors);
atomic_read       336 drivers/md/raid5-cache.c 	total_cached = atomic_read(&conf->r5c_cached_partial_stripes) +
atomic_read       337 drivers/md/raid5-cache.c 		atomic_read(&conf->r5c_cached_full_stripes);
atomic_read       348 drivers/md/raid5-cache.c 	    atomic_read(&conf->empty_inactive_list_nr) > 0)
atomic_read       365 drivers/md/raid5-cache.c 	if (atomic_read(&conf->r5c_cached_full_stripes) >=
atomic_read       407 drivers/md/raid5-cache.c 		((conf->max_degraded + 1) * atomic_read(&log->stripe_in_journal_count) +
atomic_read      1437 drivers/md/raid5-cache.c 	flushing_partial = atomic_read(&conf->r5c_flushing_partial_stripes);
atomic_read      1438 drivers/md/raid5-cache.c 	flushing_full = atomic_read(&conf->r5c_flushing_full_stripes);
atomic_read      1439 drivers/md/raid5-cache.c 	total_cached = atomic_read(&conf->r5c_cached_partial_stripes) +
atomic_read      1440 drivers/md/raid5-cache.c 		atomic_read(&conf->r5c_cached_full_stripes) -
atomic_read      1444 drivers/md/raid5-cache.c 	    atomic_read(&conf->empty_inactive_list_nr) > 0)
atomic_read      1451 drivers/md/raid5-cache.c 		 atomic_read(&conf->r5c_cached_full_stripes) - flushing_full >
atomic_read      1483 drivers/md/raid5-cache.c 			    atomic_read(&sh->count) == 0) {
atomic_read      2448 drivers/md/raid5-cache.c 		   atomic_read(&conf->active_stripes) == 0);
atomic_read      2870 drivers/md/raid5-cache.c 		BUG_ON(atomic_read(&conf->r5c_cached_partial_stripes) == 0);
atomic_read      2876 drivers/md/raid5-cache.c 		BUG_ON(atomic_read(&conf->r5c_cached_full_stripes) == 0);
atomic_read       219 drivers/md/raid5.c 	BUG_ON(atomic_read(&conf->active_stripes)==0);
atomic_read       346 drivers/md/raid5.c 		if (atomic_read(&conf->active_stripes) == 0)
atomic_read       493 drivers/md/raid5.c 	BUG_ON(atomic_read(&sh->count) != 0);
atomic_read       661 drivers/md/raid5.c 					(atomic_read(&conf->active_stripes)
atomic_read       674 drivers/md/raid5.c 			if (!atomic_read(&sh->count)) {
atomic_read       758 drivers/md/raid5.c 		if (!atomic_read(&head->count)) {
atomic_read      2443 drivers/md/raid5.c 	BUG_ON(atomic_read(&sh->count));
atomic_read      2475 drivers/md/raid5.c 		(unsigned long long)sh->sector, i, atomic_read(&sh->count),
atomic_read      2521 drivers/md/raid5.c 		if (atomic_read(&rdev->read_errors))
atomic_read      2552 drivers/md/raid5.c 		} else if (atomic_read(&rdev->read_errors)
atomic_read      2557 drivers/md/raid5.c 				    atomic_read(&rdev->read_errors),
atomic_read      2621 drivers/md/raid5.c 		(unsigned long long)sh->sector, i, atomic_read(&sh->count),
atomic_read      4714 drivers/md/raid5.c 	       atomic_read(&sh->count), sh->pd_idx, sh->qd_idx,
atomic_read      5053 drivers/md/raid5.c 		if (atomic_read(&conf->preread_active_stripes) <
atomic_read      5063 drivers/md/raid5.c 	if (atomic_read(&conf->preread_active_stripes) < IO_THRESHOLD) {
atomic_read      5111 drivers/md/raid5.c 	if (atomic_read(&conf->empty_inactive_list_nr))
atomic_read      5360 drivers/md/raid5.c 		  atomic_read(&conf->pending_full_writes), conf->bypass_count);
atomic_read      5380 drivers/md/raid5.c 		    atomic_read(&conf->pending_full_writes) == 0)) {
atomic_read      5883 drivers/md/raid5.c 			   atomic_read(&conf->reshape_stripes)==0
atomic_read      5885 drivers/md/raid5.c 		if (atomic_read(&conf->reshape_stripes) != 0)
atomic_read      5990 drivers/md/raid5.c 			   atomic_read(&conf->reshape_stripes) == 0
atomic_read      5992 drivers/md/raid5.c 		if (atomic_read(&conf->reshape_stripes) != 0)
atomic_read      6578 drivers/md/raid5.c 		return sprintf(page, "%d\n", atomic_read(&conf->active_stripes));
atomic_read      7611 drivers/md/raid5.c 		if (atomic_read(&conf->active_stripes) ||
atomic_read      7612 drivers/md/raid5.c 		    atomic_read(&conf->r5c_cached_full_stripes) ||
atomic_read      7613 drivers/md/raid5.c 		    atomic_read(&conf->r5c_cached_partial_stripes)) {
atomic_read      7631 drivers/md/raid5.c 	    atomic_read(&rdev->nr_pending)) {
atomic_read      7649 drivers/md/raid5.c 		if (atomic_read(&rdev->nr_pending)) {
atomic_read      8082 drivers/md/raid5.c 				    atomic_read(&conf->active_stripes) == 0 &&
atomic_read      8083 drivers/md/raid5.c 				    atomic_read(&conf->active_aligned_reads) == 0,
atomic_read       117 drivers/media/cec/cec-pin.c 	if (atomic_read(&pin->work_pin_num_events) < CEC_NUM_PIN_EVENTS) {
atomic_read      1042 drivers/media/cec/cec-pin.c 			atomic_read(&pin->work_irq_change) ||
atomic_read      1043 drivers/media/cec/cec-pin.c 			atomic_read(&pin->work_pin_num_events));
atomic_read      1071 drivers/media/cec/cec-pin.c 		while (atomic_read(&pin->work_pin_num_events)) {
atomic_read      1474 drivers/media/common/videobuf2/videobuf2-core.c 		       atomic_read(&q->owned_by_drv_count));
atomic_read      1487 drivers/media/common/videobuf2/videobuf2-core.c 	if (WARN_ON(atomic_read(&q->owned_by_drv_count))) {
atomic_read      1500 drivers/media/common/videobuf2/videobuf2-core.c 		WARN_ON(atomic_read(&q->owned_by_drv_count));
atomic_read      1781 drivers/media/common/videobuf2/videobuf2-core.c 		wait_event(q->done_wq, !atomic_read(&q->owned_by_drv_count));
atomic_read      1882 drivers/media/common/videobuf2/videobuf2-core.c 	if (WARN_ON(atomic_read(&q->owned_by_drv_count))) {
atomic_read      1890 drivers/media/common/videobuf2/videobuf2-core.c 		WARN_ON(atomic_read(&q->owned_by_drv_count));
atomic_read       231 drivers/media/dvb-core/dvb_ca_en50221.c 		return (atomic_read(&sl->camchange_count) != 0);
atomic_read        68 drivers/media/dvb-frontends/cxd2880/cxd2880_integ.c 	if (atomic_read(&tnr_dmd->cancel) != 0)
atomic_read       661 drivers/media/i2c/cx25840/cx25840-ir.c 	invert = (bool) atomic_read(&ir_state->rx_invert);
atomic_read       662 drivers/media/i2c/cx25840/cx25840-ir.c 	divider = (u16) atomic_read(&ir_state->rxclk_divider);
atomic_read        24 drivers/media/pci/cx18/cx18-controls.c 	if (atomic_read(&cx->ana_capturing) > 0)
atomic_read      1262 drivers/media/pci/cx18/cx18-driver.c 	if (atomic_read(&cx->tot_capturing) > 0)
atomic_read       232 drivers/media/pci/cx18/cx18-fileops.c 		if (!atomic_read(&s->q_full.depth))
atomic_read       412 drivers/media/pci/cx18/cx18-fileops.c 	if (atomic_read(&cx->ana_capturing) == 0 && s->id == -1) {
atomic_read       639 drivers/media/pci/cx18/cx18-fileops.c 	if (atomic_read(&s->q_full.depth))
atomic_read       759 drivers/media/pci/cx18/cx18-fileops.c 		if (atomic_read(&cx->ana_capturing) > 0) {
atomic_read       805 drivers/media/pci/cx18/cx18-fileops.c 			if (atomic_read(&cx->ana_capturing) > 0) {
atomic_read       851 drivers/media/pci/cx18/cx18-fileops.c 	if (atomic_read(&cx->ana_capturing)) {
atomic_read       864 drivers/media/pci/cx18/cx18-fileops.c 	if (atomic_read(&cx->ana_capturing)) {
atomic_read       273 drivers/media/pci/cx18/cx18-ioctl.c 	if (atomic_read(&cx->ana_capturing) > 0)
atomic_read       305 drivers/media/pci/cx18/cx18-ioctl.c 	if (!cx18_raw_vbi(cx) && atomic_read(&cx->ana_capturing) > 0)
atomic_read       338 drivers/media/pci/cx18/cx18-ioctl.c 	if (cx18_raw_vbi(cx) && atomic_read(&cx->ana_capturing) > 0)
atomic_read       578 drivers/media/pci/cx18/cx18-ioctl.c 	    atomic_read(&cx->ana_capturing) > 0) {
atomic_read       932 drivers/media/pci/cx18/cx18-ioctl.c 		if (!atomic_read(&cx->ana_capturing))
atomic_read       948 drivers/media/pci/cx18/cx18-ioctl.c 		if (!atomic_read(&cx->ana_capturing))
atomic_read      1034 drivers/media/pci/cx18/cx18-ioctl.c 			  atomic_read(&s->q_full.depth) * s->bufs_per_mdl * 100
atomic_read       511 drivers/media/pci/cx18/cx18-mailbox.c 		if (atomic_read(&cx->in_work_order[i].pending) == 0) {
atomic_read        57 drivers/media/pci/cx18/cx18-queue.c 	    atomic_read(&q->depth) >= CX18_MAX_FW_MDLS_PER_STREAM)
atomic_read       151 drivers/media/pci/cx18/cx18-queue.c 			if (mdl->skipped >= atomic_read(&s->q_busy.depth)-1) {
atomic_read       656 drivers/media/pci/cx18/cx18-streams.c 	if ((atomic_read(&s->q_free.depth) + atomic_read(&s->q_busy.depth)) >=
atomic_read       661 drivers/media/pci/cx18/cx18-streams.c 	if (atomic_read(&s->q_full.depth) < 2)
atomic_read       703 drivers/media/pci/cx18/cx18-streams.c 	if (atomic_read(&s->q_free.depth) == 0 ||
atomic_read       704 drivers/media/pci/cx18/cx18-streams.c 	    atomic_read(&s->q_busy.depth) >= CX18_MAX_FW_MDLS_PER_STREAM)
atomic_read       713 drivers/media/pci/cx18/cx18-streams.c 	} while (atomic_read(&s->q_busy.depth) < CX18_MAX_FW_MDLS_PER_STREAM
atomic_read       850 drivers/media/pci/cx18/cx18-streams.c 		if (atomic_read(&cx->ana_capturing) == 0)
atomic_read       902 drivers/media/pci/cx18/cx18-streams.c 	if (atomic_read(&cx->tot_capturing) == 0) {
atomic_read       931 drivers/media/pci/cx18/cx18-streams.c 		if (atomic_read(&cx->tot_capturing) == 0) {
atomic_read       972 drivers/media/pci/cx18/cx18-streams.c 	if (atomic_read(&cx->tot_capturing) == 0)
atomic_read       999 drivers/media/pci/cx18/cx18-streams.c 	if (atomic_read(&cx->tot_capturing) > 0)
atomic_read       477 drivers/media/pci/cx23885/cx23885-alsa.c 	count = atomic_read(&chip->count);
atomic_read       109 drivers/media/pci/cx23885/cx23885-input.c 		params.shutdown = atomic_read(&dev->ir_input_stopping);
atomic_read       120 drivers/media/pci/cx23885/cx23885-input.c 		params.shutdown = atomic_read(&dev->ir_input_stopping);
atomic_read       653 drivers/media/pci/cx23885/cx23888-ir.c 	bool invert = (bool) atomic_read(&state->rx_invert);
atomic_read       654 drivers/media/pci/cx23885/cx23888-ir.c 	u16 divider = (u16) atomic_read(&state->rxclk_divider);
atomic_read       620 drivers/media/pci/cx25821/cx25821-alsa.c 	count = atomic_read(&chip->count);
atomic_read       563 drivers/media/pci/cx88/cx88-alsa.c 	count = atomic_read(&chip->count);
atomic_read       576 drivers/media/pci/intel/ipu3/ipu3-cio2.c 			b->vbb.sequence = atomic_read(&q->frame_sequence);
atomic_read       604 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		.u.frame_sync.frame_sequence = atomic_read(&q->frame_sequence),
atomic_read       106 drivers/media/pci/ivtv/ivtv-controls.c 	if (atomic_read(&itv->decoding)) {
atomic_read      1405 drivers/media/pci/ivtv/ivtv-driver.c 		if (atomic_read(&itv->capturing) > 0)
atomic_read      1414 drivers/media/pci/ivtv/ivtv-driver.c 		if (atomic_read(&itv->decoding) > 0) {
atomic_read       352 drivers/media/pci/ivtv/ivtv-fileops.c 	if (atomic_read(&itv->capturing) == 0 && s->fh == NULL) {
atomic_read       517 drivers/media/pci/ivtv/ivtv-fileops.c 	if (atomic_read(&itv->decoding) == 0) {
atomic_read       892 drivers/media/pci/ivtv/ivtv-fileops.c 		if (atomic_read(&itv->capturing) > 0) {
atomic_read       996 drivers/media/pci/ivtv/ivtv-fileops.c 			if (atomic_read(&itv->capturing) > 0) {
atomic_read      1048 drivers/media/pci/ivtv/ivtv-fileops.c 	if (atomic_read(&itv->capturing))
atomic_read      1055 drivers/media/pci/ivtv/ivtv-fileops.c 	if (atomic_read(&itv->capturing)) {
atomic_read       334 drivers/media/pci/ivtv/ivtv-firmware.c 	if (!res && !atomic_read(&itv->capturing) &&
atomic_read       335 drivers/media/pci/ivtv/ivtv-firmware.c 	    (!atomic_read(&itv->decoding) ||
atomic_read       336 drivers/media/pci/ivtv/ivtv-firmware.c 	     (atomic_read(&itv->decoding) < 2 && test_bit(IVTV_F_I_DEC_YUV,
atomic_read       365 drivers/media/pci/ivtv/ivtv-firmware.c 	if (res && !atomic_read(&itv->capturing) &&
atomic_read       366 drivers/media/pci/ivtv/ivtv-firmware.c 						!atomic_read(&itv->decoding)) {
atomic_read       163 drivers/media/pci/ivtv/ivtv-ioctl.c 	if (atomic_read(&itv->decoding) > 0) {
atomic_read       272 drivers/media/pci/ivtv/ivtv-ioctl.c 		if (atomic_read(&itv->decoding) == 0)
atomic_read       283 drivers/media/pci/ivtv/ivtv-ioctl.c 		if (!atomic_read(&itv->decoding))
atomic_read       287 drivers/media/pci/ivtv/ivtv-ioctl.c 		if (atomic_read(&itv->decoding) > 0) {
atomic_read       297 drivers/media/pci/ivtv/ivtv-ioctl.c 		if (!atomic_read(&itv->decoding))
atomic_read       589 drivers/media/pci/ivtv/ivtv-ioctl.c 	if (atomic_read(&itv->capturing) > 0)
atomic_read       607 drivers/media/pci/ivtv/ivtv-ioctl.c 	if (!ivtv_raw_vbi(itv) && atomic_read(&itv->capturing) > 0)
atomic_read       626 drivers/media/pci/ivtv/ivtv-ioctl.c 	if (ivtv_raw_vbi(itv) && atomic_read(&itv->capturing) > 0)
atomic_read       995 drivers/media/pci/ivtv/ivtv-ioctl.c 	if (atomic_read(&itv->capturing) > 0) {
atomic_read      1173 drivers/media/pci/ivtv/ivtv-ioctl.c 	    atomic_read(&itv->capturing) > 0 ||
atomic_read      1174 drivers/media/pci/ivtv/ivtv-ioctl.c 	    atomic_read(&itv->decoding) > 0) {
atomic_read      1267 drivers/media/pci/ivtv/ivtv-ioctl.c 	if (!atomic_read(&itv->capturing))
atomic_read      1302 drivers/media/pci/ivtv/ivtv-ioctl.c 		if (!atomic_read(&itv->capturing))
atomic_read      1315 drivers/media/pci/ivtv/ivtv-ioctl.c 		if (!atomic_read(&itv->capturing))
atomic_read       798 drivers/media/pci/ivtv/ivtv-irq.c 		if (atomic_read(&itv->yuv_info.next_dma_frame) >= 0)
atomic_read       833 drivers/media/pci/ivtv/ivtv-irq.c 	int last_dma_frame = atomic_read(&yi->next_dma_frame);
atomic_read       844 drivers/media/pci/ivtv/ivtv-irq.c 			if (next_dma_frame >= 0 && next_dma_frame != atomic_read(&yi->next_fill_frame)) {
atomic_read       898 drivers/media/pci/ivtv/ivtv-irq.c 					(u8)(atomic_read(&yi->next_dma_frame) -
atomic_read       549 drivers/media/pci/ivtv/ivtv-streams.c 	if (atomic_read(&itv->capturing) == 0) {
atomic_read       617 drivers/media/pci/ivtv/ivtv-streams.c 	if (atomic_read(&itv->capturing) == 0) {
atomic_read       808 drivers/media/pci/ivtv/ivtv-streams.c 	if (atomic_read(&itv->capturing) == 0)
atomic_read       887 drivers/media/pci/ivtv/ivtv-streams.c 	if (atomic_read(&itv->capturing) > 0) {
atomic_read      1013 drivers/media/pci/ivtv/ivtv-streams.c 		if (atomic_read(&itv->capturing) == 0) {
atomic_read      1036 drivers/media/pci/ivtv/ivtv-streams.c 	if (atomic_read(&itv->capturing) == 0)
atomic_read       947 drivers/media/pci/ivtv/ivtv-yuv.c 	if (atomic_read(&yi->next_dma_frame) == -1)
atomic_read       950 drivers/media/pci/ivtv/ivtv-yuv.c 	draw = atomic_read(&yi->next_fill_frame);
atomic_read       951 drivers/media/pci/ivtv/ivtv-yuv.c 	display = atomic_read(&yi->next_dma_frame);
atomic_read        66 drivers/media/platform/exynos4-is/fimc-lite-reg.c 	if (atomic_read(&dev->out_path) == FIMC_IO_DMA) {
atomic_read       270 drivers/media/platform/exynos4-is/fimc-lite.c 	if (atomic_read(&fimc->out_path) != FIMC_IO_DMA)
atomic_read       465 drivers/media/platform/exynos4-is/fimc-lite.c 	if (atomic_read(&fimc->out_path) != FIMC_IO_DMA) {
atomic_read       480 drivers/media/platform/exynos4-is/fimc-lite.c 	    atomic_read(&fimc->out_path) != FIMC_IO_DMA)
atomic_read       515 drivers/media/platform/exynos4-is/fimc-lite.c 	    atomic_read(&fimc->out_path) == FIMC_IO_DMA) {
atomic_read      1084 drivers/media/platform/exynos4-is/fimc-lite.c 	if ((atomic_read(&fimc->out_path) == FIMC_IO_ISP &&
atomic_read      1086 drivers/media/platform/exynos4-is/fimc-lite.c 	    (atomic_read(&fimc->out_path) == FIMC_IO_DMA &&
atomic_read      1213 drivers/media/platform/exynos4-is/fimc-lite.c 	if (atomic_read(&fimc->out_path) != FIMC_IO_ISP)
atomic_read      1583 drivers/media/platform/exynos4-is/fimc-lite.c 	fimc_lite_hw_init(fimc, atomic_read(&fimc->out_path) == FIMC_IO_ISP);
atomic_read       155 drivers/media/platform/m2m-deinterlace.c 	    !atomic_read(&ctx->dev->busy)) {
atomic_read      1222 drivers/media/platform/omap3isp/isp.c 		if (!wait_event_timeout(*wait, !atomic_read(stopping),
atomic_read      1467 drivers/media/platform/omap3isp/ispccdc.c 	event.u.frame_sync.frame_sequence = atomic_read(&pipe->frame_number);
atomic_read       275 drivers/media/platform/omap3isp/isphist.c 	if (atomic_read(&hist->buf_err) || hist->state != ISPSTAT_ENABLED) {
atomic_read       918 drivers/media/platform/omap3isp/ispstat.c 		stat->frame_number = atomic_read(&pipe->frame_number);
atomic_read       544 drivers/media/platform/omap3isp/ispvideo.c 		buf->vb.sequence = atomic_read(&pipe->frame_number);
atomic_read       978 drivers/media/platform/qcom/camss/camss.c 	if (atomic_read(&camss->ref_count) == 0)
atomic_read       103 drivers/media/platform/qcom/venus/hfi.c 			       !atomic_read(&core->insts_count));
atomic_read       150 drivers/media/platform/s5p-mfc/s5p_mfc.c 	if (atomic_read(&dev->watchdog_cnt) >= MFC_WATCHDOG_CNT) {
atomic_read        64 drivers/media/platform/s5p-mfc/s5p_mfc_pm.c 	mfc_debug(3, "+ %d\n", atomic_read(&clk_ref));
atomic_read        72 drivers/media/platform/s5p-mfc/s5p_mfc_pm.c 	mfc_debug(3, "- %d\n", atomic_read(&clk_ref));
atomic_read       170 drivers/media/platform/sti/c8sectpfe/c8sectpfe-core.c 	if (!atomic_read(&fei->fw_loaded)) {
atomic_read       267 drivers/media/platform/sti/c8sectpfe/c8sectpfe-core.c 	if (!atomic_read(&fei->fw_loaded)) {
atomic_read       165 drivers/media/radio/radio-shark.c 		brightness = atomic_read(&shark->brightness[i]);
atomic_read       152 drivers/media/radio/radio-shark2.c 		brightness = atomic_read(&shark->brightness[i]);
atomic_read      1084 drivers/media/radio/radio-si476x.c 	    atomic_read(&radio->core->is_alive))
atomic_read      1109 drivers/media/radio/radio-si476x.c 						 !atomic_read(&radio->core->is_alive)));
atomic_read      1113 drivers/media/radio/radio-si476x.c 		if (!atomic_read(&radio->core->is_alive))
atomic_read      1140 drivers/media/radio/radio-si476x.c 		if (atomic_read(&radio->core->is_alive))
atomic_read      1143 drivers/media/radio/radio-si476x.c 		if (!atomic_read(&radio->core->is_alive))
atomic_read       324 drivers/media/radio/wl128x/fmdrv_common.c 		if (num_fm_hci_cmds && atomic_read(&fmdev->tx_cnt))
atomic_read       339 drivers/media/radio/wl128x/fmdrv_common.c 	if (!atomic_read(&fmdev->tx_cnt))
atomic_read       110 drivers/media/usb/cx231xx/cx231xx-audio.c 	if (atomic_read(&dev->stream_started) == 0)
atomic_read       201 drivers/media/usb/cx231xx/cx231xx-audio.c 	if (atomic_read(&dev->stream_started) == 0)
atomic_read       500 drivers/media/usb/cx231xx/cx231xx-audio.c 		if (atomic_read(&dev->stream_started) > 0) {
atomic_read       538 drivers/media/usb/cx231xx/cx231xx-audio.c 	if (atomic_read(&dev->stream_started) > 0) {
atomic_read       560 drivers/media/usb/cx231xx/cx231xx-audio.c 	if (atomic_read(&dev->stream_started)) {
atomic_read        65 drivers/media/usb/cx231xx/cx231xx-core.c 	if (atomic_read(&dev->devlist_count) > 0) {
atomic_read      1897 drivers/media/usb/cx231xx/cx231xx-video.c 				if (atomic_read(&dev->devlist_count) > 0) {
atomic_read       110 drivers/media/usb/em28xx/em28xx-audio.c 	if (atomic_read(&dev->adev.stream_started) == 0)
atomic_read       339 drivers/media/usb/em28xx/em28xx-audio.c 	if (atomic_read(&dev->adev.stream_started) > 0) {
atomic_read       394 drivers/media/usb/em28xx/em28xx-audio.c 	if (atomic_read(&adev->stream_started) > 0) {
atomic_read       421 drivers/media/usb/em28xx/em28xx-audio.c 	if (atomic_read(&adev->stream_started)) {
atomic_read      1270 drivers/media/usb/gspca/cpia1.c 	framerate = atomic_read(&sd->fps);
atomic_read      1393 drivers/media/usb/gspca/cpia1.c 	cam_exposure = atomic_read(&sd->cam_exposure);
atomic_read       343 drivers/media/usb/gspca/pac207.c 	int avg_lum = atomic_read(&sd->avg_lum);
atomic_read       727 drivers/media/usb/gspca/pac7302.c 	int avg_lum = atomic_read(&sd->avg_lum);
atomic_read       492 drivers/media/usb/gspca/pac7311.c 	int avg_lum = atomic_read(&sd->avg_lum);
atomic_read      2142 drivers/media/usb/gspca/sn9c20x.c 	avg_lum = atomic_read(&sd->avg_lum);
atomic_read       880 drivers/media/usb/gspca/sonixb.c 	avg_lum = atomic_read(&sd->avg_lum);
atomic_read      2584 drivers/media/usb/gspca/sonixj.c 	delta = atomic_read(&sd->avg_lum);
atomic_read      1416 drivers/media/usb/s2255/s2255drv.c 	state = atomic_read(&dev->fw_data->fw_state);
atomic_read      1425 drivers/media/usb/s2255/s2255drv.c 				   ((atomic_read(&dev->fw_data->fw_state)
atomic_read      1427 drivers/media/usb/s2255/s2255drv.c 				    (atomic_read(&dev->fw_data->fw_state)
atomic_read      1431 drivers/media/usb/s2255/s2255drv.c 		state = atomic_read(&dev->fw_data->fw_state);
atomic_read      1439 drivers/media/usb/s2255/s2255drv.c 				   ((atomic_read(&dev->fw_data->fw_state)
atomic_read      1441 drivers/media/usb/s2255/s2255drv.c 				    (atomic_read(&dev->fw_data->fw_state)
atomic_read      1445 drivers/media/usb/s2255/s2255drv.c 		state = atomic_read(&dev->fw_data->fw_state);
atomic_read      1555 drivers/media/usb/s2255/s2255drv.c 		atomic_read(&dev->num_channels));
atomic_read      1672 drivers/media/usb/s2255/s2255drv.c 	if (atomic_read(&dev->num_channels) == 0) {
atomic_read      1676 drivers/media/usb/s2255/s2255drv.c 	if (atomic_read(&dev->num_channels) != MAX_CHANNELS)
atomic_read      2345 drivers/media/usb/s2255/s2255drv.c 	int channels = atomic_read(&dev->num_channels);
atomic_read       506 drivers/media/usb/stkwebcam/stk-webcam.c 			if (atomic_read(&dev->urbs_used) && is_present(dev))
atomic_read       193 drivers/media/usb/tm6000/tm6000-alsa.c 	if (atomic_read(&core->stream_started) > 0) {
atomic_read       210 drivers/media/usb/tm6000/tm6000-alsa.c 	if (atomic_read(&core->stream_started) == 0)
atomic_read       296 drivers/media/usb/tm6000/tm6000-alsa.c 	if (atomic_read(&core->stream_started) > 0) {
atomic_read       328 drivers/media/usb/tm6000/tm6000-alsa.c 	if (atomic_read(&core->stream_started)) {
atomic_read        80 drivers/media/usb/usbtv/usbtv-audio.c 	if (atomic_read(&chip->snd_stream)) {
atomic_read       146 drivers/media/usb/usbtv/usbtv-audio.c 	if (!atomic_read(&chip->snd_stream))
atomic_read       282 drivers/media/usb/usbtv/usbtv-audio.c 	if (atomic_read(&usbtv->snd_stream) && usbtv->snd_bulk_urb)
atomic_read       288 drivers/media/usb/usbtv/usbtv-audio.c 	if (atomic_read(&usbtv->snd_stream) && usbtv->snd_bulk_urb)
atomic_read       299 drivers/media/usb/usbtv/usbtv-audio.c 	if (atomic_read(&chip->snd_stream))
atomic_read       270 drivers/media/v4l2-core/v4l2-clk.c 	if (WARN(atomic_read(&clk->use_count),
atomic_read       287 drivers/media/v4l2-core/v4l2-dev.c 	if (atomic_read(&global->prios[V4L2_PRIORITY_RECORD]) > 0)
atomic_read       289 drivers/media/v4l2-core/v4l2-dev.c 	if (atomic_read(&global->prios[V4L2_PRIORITY_INTERACTIVE]) > 0)
atomic_read       291 drivers/media/v4l2-core/v4l2-dev.c 	if (atomic_read(&global->prios[V4L2_PRIORITY_BACKGROUND]) > 0)
atomic_read       508 drivers/message/fusion/mptlan.c 		  priv->total_posted,atomic_read(&priv->buckets_out)));
atomic_read       515 drivers/message/fusion/mptlan.c 	while (atomic_read(&priv->buckets_out) && time_before(jiffies, timeout))
atomic_read       835 drivers/message/fusion/mptlan.c 		 atomic_read(&priv->buckets_out)));
atomic_read       837 drivers/message/fusion/mptlan.c 	if (atomic_read(&priv->buckets_out) < priv->bucketthresh)
atomic_read       842 drivers/message/fusion/mptlan.c 		  atomic_read(&priv->buckets_out), priv->total_received));
atomic_read       953 drivers/message/fusion/mptlan.c /**/		  atomic_read(&priv->buckets_out), priv->total_received));
atomic_read      1110 drivers/message/fusion/mptlan.c 			atomic_read(&priv->buckets_out));
atomic_read      1115 drivers/message/fusion/mptlan.c 			remaining, atomic_read(&priv->buckets_out));
atomic_read      1118 drivers/message/fusion/mptlan.c 	    ((atomic_read(&priv->buckets_out) - remaining) >
atomic_read      1154 drivers/message/fusion/mptlan.c 	curr = atomic_read(&priv->buckets_out);
atomic_read      1280 drivers/message/fusion/mptlan.c 		  __func__, buckets, atomic_read(&priv->buckets_out)));
atomic_read      3759 drivers/message/fusion/mptsas.c 						atomic_read(&sdev->device_busy)));
atomic_read       606 drivers/mfd/ab8500-core.c 	if (atomic_read(&ab8500->transfer_ongoing))
atomic_read      2344 drivers/mfd/db8500-prcmu.c 	return (atomic_read(&ac_wake_req_state) != 0);
atomic_read        57 drivers/mfd/mfd-core.c 	WARN_ON(atomic_read(cell->usage_count) < 0);
atomic_read       299 drivers/mfd/si476x-cmd.c 				atomic_read(&core->cts),
atomic_read       314 drivers/mfd/si476x-cmd.c 					atomic_read(&core->cts),
atomic_read       385 drivers/mfd/si476x-cmd.c 				    atomic_read(&core->stc));
atomic_read       521 drivers/mfd/si476x-i2c.c 	if (atomic_read(&core->is_alive))
atomic_read       128 drivers/mfd/twl6030-irq.c 		chained_wakeups = atomic_read(&pdata->wakeirqs);
atomic_read       290 drivers/misc/cb710/core.c 	BUG_ON(atomic_read(&chip->slot_refs_count) != 0);
atomic_read       303 drivers/misc/cb710/core.c 	BUG_ON(atomic_read(&chip->slot_refs_count) != 0);
atomic_read       358 drivers/misc/cxl/file.c 	if (ctx->afu_driver_ops && atomic_read(&ctx->afu_driver_events))
atomic_read       474 drivers/misc/cxl/file.c 	if (ctx->afu_driver_ops && atomic_read(&ctx->afu_driver_events)) {
atomic_read      1126 drivers/misc/cxl/pci.c 	if (atomic_read(&afu->configured_state) != -1) {
atomic_read      1962 drivers/misc/cxl/pci.c 			 atomic_read(&adapter->contexts_num));
atomic_read       100 drivers/misc/habanalabs/command_buffer.c 	if ((hdev->disabled) || ((atomic_read(&hdev->in_reset)) &&
atomic_read       220 drivers/misc/habanalabs/command_buffer.c 			atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
atomic_read       624 drivers/misc/habanalabs/command_submission.c 			atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
atomic_read        21 drivers/misc/habanalabs/device.c 	if ((hdev->disabled) || (atomic_read(&hdev->in_reset)))
atomic_read        33 drivers/misc/habanalabs/device.c 	else if (atomic_read(&hdev->in_reset))
atomic_read       249 drivers/misc/habanalabs/habanalabs_ioctl.c 			atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
atomic_read       302 drivers/misc/habanalabs/habanalabs_ioctl.c 			atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
atomic_read       313 drivers/misc/habanalabs/memory.c 		if (atomic_read(&phys_pg_pack->mapping_cnt) > 0) {
atomic_read      1030 drivers/misc/habanalabs/memory.c 	if (atomic_read(&phys_pg_pack->mapping_cnt) == 0) {
atomic_read      1163 drivers/misc/habanalabs/memory.c 			atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
atomic_read       250 drivers/misc/habanalabs/sysfs.c 	if (atomic_read(&hdev->in_reset))
atomic_read        47 drivers/misc/ibmasm/command.c 	dbg("command count: %d\n", atomic_read(&command_count));
atomic_read        58 drivers/misc/ibmasm/command.c 	dbg("command count: %d\n", atomic_read(&command_count));
atomic_read       493 drivers/misc/lis3lv02d/lis3lv02d.c 	if (atomic_read(&lis3->wake_thread))
atomic_read       647 drivers/misc/lis3lv02d/lis3lv02d.c 	if (atomic_read(&lis3->count))
atomic_read       170 drivers/misc/mic/scif/scif_dma.c 	if ((atomic_read(&ep->rma_info.tcw_total_pages)
atomic_read       176 drivers/misc/mic/scif/scif_dma.c 			 atomic_read(&ep->rma_info.tcw_total_pages),
atomic_read       194 drivers/misc/mic/scif/scif_nm.c 				 (atomic_read(&scifdev->disconn_rescnt)
atomic_read       209 drivers/misc/mic/scif/scif_nm.c 				   (atomic_read(&scifdev->disconn_rescnt) == 1),
atomic_read        65 drivers/misc/mic/scif/scif_rma.c 	    !atomic_read(&ep->rma_info.tw_refcount) &&
atomic_read        66 drivers/misc/mic/scif/scif_rma.c 	    !atomic_read(&ep->rma_info.tcw_refcount) &&
atomic_read        67 drivers/misc/mic/scif/scif_rma.c 	    !atomic_read(&ep->rma_info.fence_refcount))
atomic_read       260 drivers/misc/phantom.c 	pr_debug("phantom_poll: %d\n", atomic_read(&dev->counter));
atomic_read       265 drivers/misc/phantom.c 	else if (atomic_read(&dev->counter))
atomic_read       268 drivers/misc/phantom.c 	pr_debug("phantom_poll end: %x/%d\n", mask, atomic_read(&dev->counter));
atomic_read       410 drivers/misc/sgi-gru/grufault.c 	if (atomic_read(&gts->ts_gms->ms_range_active))
atomic_read       623 drivers/misc/sgi-gru/grufault.c 			   atomic_read(&gms->ms_range_active) == 0);
atomic_read       218 drivers/misc/sgi-gru/grutlbpurge.c 		range->start, range->end, atomic_read(&gms->ms_range_active));
atomic_read        94 drivers/misc/sgi-xp/xpc_channel.c 	if (atomic_read(&ch->kthreads_assigned) > 0 ||
atomic_read        95 drivers/misc/sgi-xp/xpc_channel.c 	    atomic_read(&ch->references) > 0) {
atomic_read       123 drivers/misc/sgi-xp/xpc_channel.c 	if (atomic_read(&ch->n_to_notify) > 0) {
atomic_read       136 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(atomic_read(&ch->n_to_notify) != 0);
atomic_read       493 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(atomic_read(&ch->kthreads_assigned) != 0);
atomic_read       494 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(atomic_read(&ch->kthreads_idle) != 0);
atomic_read       495 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(atomic_read(&ch->kthreads_active) != 0);
atomic_read       784 drivers/misc/sgi-xp/xpc_channel.c 	if (atomic_read(&ch->kthreads_idle) > 0) {
atomic_read       794 drivers/misc/sgi-xp/xpc_channel.c 	if (atomic_read(&ch->n_on_msg_allocate_wq) > 0)
atomic_read       342 drivers/misc/sgi-xp/xpc_main.c 	       atomic_read(&part->nchannels_active) > 0 ||
atomic_read       362 drivers/misc/sgi-xp/xpc_main.c 				(atomic_read(&part->channel_mgr_requests) > 0 ||
atomic_read       365 drivers/misc/sgi-xp/xpc_main.c 				 atomic_read(&part->nchannels_active) == 0 &&
atomic_read       494 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(atomic_read(&part->nchannels_engaged) != 0);
atomic_read       495 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(atomic_read(&part->nchannels_active) != 0);
atomic_read       505 drivers/misc/sgi-xp/xpc_main.c 	wait_event(part->teardown_wq, (atomic_read(&part->references) == 0));
atomic_read       612 drivers/misc/sgi-xp/xpc_main.c 	int idle = atomic_read(&ch->kthreads_idle);
atomic_read       613 drivers/misc/sgi-xp/xpc_main.c 	int assigned = atomic_read(&ch->kthreads_assigned);
atomic_read       829 drivers/misc/sgi-xp/xpc_main.c 			if (atomic_read(&ch->kthreads_assigned) <
atomic_read      1364 drivers/misc/sgi-xp/xpc_uv.c 	if (atomic_read(&ch->n_on_msg_allocate_wq) > 0)
atomic_read      1470 drivers/misc/sgi-xp/xpc_uv.c 		if (atomic_read(&ch->kthreads_idle) > 0)
atomic_read      1634 drivers/misc/sgi-xp/xpc_uv.c 		if (atomic_read(&ch->n_to_notify) == 0)
atomic_read       362 drivers/misc/vmw_vmci/vmci_doorbell.c 		    atomic_read(&dbell->active) == 1) {
atomic_read        75 drivers/misc/vmw_vmci/vmci_guest.c 	return atomic_read(&vmci_num_guest_devices) != 0;
atomic_read       108 drivers/misc/vmw_vmci/vmci_host.c 	     atomic_read(&vmci_host_active_users) > 0);
atomic_read       811 drivers/mmc/core/core.c 		stop = abort ? atomic_read(abort) : 0;
atomic_read      1061 drivers/mmc/core/sdio.c 	if (atomic_read(&card->sdio_funcs_probed) > 1) {
atomic_read       147 drivers/mtd/maps/vmu-flash.c 		if (atomic_read(&mdev->busy) == 1) {
atomic_read       149 drivers/mtd/maps/vmu-flash.c 				atomic_read(&mdev->busy) == 0, HZ);
atomic_read       150 drivers/mtd/maps/vmu-flash.c 			if (atomic_read(&mdev->busy) == 1) {
atomic_read       173 drivers/mtd/maps/vmu-flash.c 			(atomic_read(&mdev->busy) == 0 ||
atomic_read       174 drivers/mtd/maps/vmu-flash.c 			atomic_read(&mdev->busy) == 2), HZ * 3);
atomic_read       180 drivers/mtd/maps/vmu-flash.c 		if (error || atomic_read(&mdev->busy) == 2) {
atomic_read       181 drivers/mtd/maps/vmu-flash.c 			if (atomic_read(&mdev->busy) == 2)
atomic_read       252 drivers/mtd/maps/vmu-flash.c 		if (atomic_read(&mdev->busy) == 1) {
atomic_read       254 drivers/mtd/maps/vmu-flash.c 				atomic_read(&mdev->busy) == 0, HZ);
atomic_read       255 drivers/mtd/maps/vmu-flash.c 			if (atomic_read(&mdev->busy) == 1) {
atomic_read       268 drivers/mtd/maps/vmu-flash.c 			atomic_read(&mdev->busy) == 0, HZ/10);
atomic_read       274 drivers/mtd/maps/vmu-flash.c 		if (atomic_read(&mdev->busy) == 2) {
atomic_read       655 drivers/mtd/maps/vmu-flash.c 	if (atomic_read(&mdev->busy) == 1) {
atomic_read       657 drivers/mtd/maps/vmu-flash.c 			atomic_read(&mdev->busy) == 0, HZ);
atomic_read       658 drivers/mtd/maps/vmu-flash.c 		if (atomic_read(&mdev->busy) == 1) {
atomic_read      4966 drivers/net/bonding/bond_main.c 	WARN_ON(atomic_read(&netpoll_block_tx));
atomic_read       475 drivers/net/can/c_can/c_can.c 	idx = fls(atomic_read(&priv->tx_active));
atomic_read       829 drivers/net/can/usb/ems_usb.c 		if (atomic_read(&dev->active_tx_urbs) >= MAX_TX_URBS ||
atomic_read       790 drivers/net/can/usb/esd_usb2.c 	if (atomic_read(&priv->active_tx_jobs) >= MAX_TX_URBS)
atomic_read       545 drivers/net/can/usb/gs_usb.c 		if (atomic_read(&dev->active_tx_urbs) >= GS_MAX_TX_URBS)
atomic_read       198 drivers/net/can/usb/mcba_usb.c 	if (!atomic_read(&priv->free_ctx_cnt))
atomic_read       357 drivers/net/can/usb/peak_usb/pcan_usb_core.c 		if (atomic_read(&dev->active_tx_urbs) >= PCAN_USB_MAX_TX_URBS)
atomic_read       674 drivers/net/can/usb/usb_8dev.c 	else if (atomic_read(&priv->active_tx_urbs) >= MAX_TX_URBS)
atomic_read      1576 drivers/net/ethernet/alteon/acenic.c 	cur_size = atomic_read(&ap->cur_rx_bufs);
atomic_read      1586 drivers/net/ethernet/alteon/acenic.c 		cur_size = atomic_read(&ap->cur_mini_bufs);
atomic_read      1597 drivers/net/ethernet/alteon/acenic.c 	cur_size = atomic_read(&ap->cur_jumbo_bufs);
atomic_read      2173 drivers/net/ethernet/alteon/acenic.c 		cur_size = atomic_read(&ap->cur_rx_bufs);
atomic_read      2187 drivers/net/ethernet/alteon/acenic.c 			cur_size = atomic_read(&ap->cur_mini_bufs);
atomic_read      2204 drivers/net/ethernet/alteon/acenic.c 			cur_size = atomic_read(&ap->cur_jumbo_bufs);
atomic_read       243 drivers/net/ethernet/amazon/ena/ena_com.c 	cnt = (u16)atomic_read(&admin_queue->outstanding_cmds);
atomic_read      1459 drivers/net/ethernet/amazon/ena/ena_com.c 	while (atomic_read(&admin_queue->outstanding_cmds) != 0) {
atomic_read        19 drivers/net/ethernet/aquantia/atlantic/aq_utils.h 		flags_old = atomic_read(flags);
atomic_read        29 drivers/net/ethernet/aquantia/atlantic/aq_utils.h 		flags_old = atomic_read(flags);
atomic_read        36 drivers/net/ethernet/aquantia/atlantic/aq_utils.h 	return atomic_read(flags) & mask;
atomic_read       553 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c 		pmbox->stats.dpc = atomic_read(&self->dpc);
atomic_read      1542 drivers/net/ethernet/atheros/atl1c/atl1c_main.c 	u16 next_to_clean = atomic_read(&tpd_ring->next_to_clean);
atomic_read      1914 drivers/net/ethernet/atheros/atl1c/atl1c_main.c 	next_to_clean = atomic_read(&tpd_ring->next_to_clean);
atomic_read      1234 drivers/net/ethernet/atheros/atl1e/atl1e_main.c 	u16 next_to_clean = atomic_read(&tx_ring->next_to_clean);
atomic_read      1558 drivers/net/ethernet/atheros/atl1e/atl1e_main.c 	next_to_clean = atomic_read(&tx_ring->next_to_clean);
atomic_read      1499 drivers/net/ethernet/atheros/atlx/atl1.c 	value = ((atomic_read(&adapter->tpd_ring.next_to_use)
atomic_read      1501 drivers/net/ethernet/atheros/atlx/atl1.c 		((atomic_read(&adapter->rrd_ring.next_to_clean)
atomic_read      1503 drivers/net/ethernet/atheros/atlx/atl1.c 		((atomic_read(&adapter->rfd_ring.next_to_use)
atomic_read      1745 drivers/net/ethernet/atheros/atlx/atl1.c 	tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use);
atomic_read      1746 drivers/net/ethernet/atheros/atlx/atl1.c 	rfd_next_to_use = atomic_read(&adapter->rfd_ring.next_to_use);
atomic_read      1747 drivers/net/ethernet/atheros/atlx/atl1.c 	rrd_next_to_clean = atomic_read(&adapter->rrd_ring.next_to_clean);
atomic_read      1844 drivers/net/ethernet/atheros/atlx/atl1.c 	next_next = rfd_next_to_use = atomic_read(&rfd_ring->next_to_use);
atomic_read      1915 drivers/net/ethernet/atheros/atlx/atl1.c 	rrd_next_to_clean = atomic_read(&rrd_ring->next_to_clean);
atomic_read      2035 drivers/net/ethernet/atheros/atlx/atl1.c 		tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use);
atomic_read      2037 drivers/net/ethernet/atheros/atlx/atl1.c 		    atomic_read(&adapter->rfd_ring.next_to_use);
atomic_read      2039 drivers/net/ethernet/atheros/atlx/atl1.c 		    atomic_read(&adapter->rrd_ring.next_to_clean);
atomic_read      2061 drivers/net/ethernet/atheros/atlx/atl1.c 	sw_tpd_next_to_clean = atomic_read(&tpd_ring->next_to_clean);
atomic_read      2093 drivers/net/ethernet/atheros/atlx/atl1.c 	u16 next_to_clean = atomic_read(&tpd_ring->next_to_clean);
atomic_read      2094 drivers/net/ethernet/atheros/atlx/atl1.c 	u16 next_to_use = atomic_read(&tpd_ring->next_to_use);
atomic_read      2202 drivers/net/ethernet/atheros/atlx/atl1.c 	next_to_use = atomic_read(&tpd_ring->next_to_use);
atomic_read      2295 drivers/net/ethernet/atheros/atlx/atl1.c 	u16 next_to_use = (u16) atomic_read(&tpd_ring->next_to_use);
atomic_read      2394 drivers/net/ethernet/atheros/atlx/atl1.c 		(u16) atomic_read(&tpd_ring->next_to_use));
atomic_read       478 drivers/net/ethernet/atheros/atlx/atl2.c 		txs_write_ptr = (u32) atomic_read(&adapter->txs_write_ptr);
atomic_read       490 drivers/net/ethernet/atheros/atlx/atl2.c 		txd_read_ptr = (u32) atomic_read(&adapter->txd_read_ptr);
atomic_read       808 drivers/net/ethernet/atheros/atlx/atl2.c 	u32 txs_write_ptr = (u32) atomic_read(&adapter->txs_write_ptr);
atomic_read       818 drivers/net/ethernet/atheros/atlx/atl2.c 	u32 txd_read_ptr = (u32)atomic_read(&adapter->txd_read_ptr);
atomic_read       399 drivers/net/ethernet/aurora/nb8800.c 	if (atomic_read(&priv->tx_free) <= NB8800_DESC_LOW) {
atomic_read      3327 drivers/net/ethernet/broadcom/bnx2.c 	if (unlikely(atomic_read(&bp->intr_sem) != 0))
atomic_read      3344 drivers/net/ethernet/broadcom/bnx2.c 	if (unlikely(atomic_read(&bp->intr_sem) != 0))
atomic_read      3380 drivers/net/ethernet/broadcom/bnx2.c 	if (unlikely(atomic_read(&bp->intr_sem) != 0))
atomic_read      6192 drivers/net/ethernet/broadcom/bnx2.c 	if (atomic_read(&bp->intr_sem) != 0)
atomic_read      6561 drivers/net/ethernet/broadcom/bnx2.c 		   atomic_read(&bp->intr_sem), val1);
atomic_read      3182 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		if (atomic_read(&bp->pdev->enable_cnt) != 1)
atomic_read      1876 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	DP(BNX2X_MSG_SP, "bp->cq_spq_left %x\n", atomic_read(&bp->cq_spq_left));
atomic_read      3876 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		if (!atomic_read(&bp->eq_spq_left)) {
atomic_read      3882 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	} else if (!atomic_read(&bp->cq_spq_left)) {
atomic_read      3930 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	   atomic_read(&bp->cq_spq_left), atomic_read(&bp->eq_spq_left));
atomic_read      5470 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 			hw_cons, sw_cons, atomic_read(&bp->eq_spq_left));
atomic_read      5670 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	if (atomic_read(&bp->interrupt_occurred)) {
atomic_read      13221 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	if (atomic_read(&pdev->enable_cnt) == 1) {
atomic_read      13392 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	if (atomic_read(&pdev->enable_cnt) == 1)
atomic_read      14146 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	if (atomic_read(&pdev->enable_cnt) == 1)
atomic_read      14238 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		if (atomic_read(&pdev->enable_cnt) == 1)
atomic_read      14596 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 			if (!atomic_read(&bp->cq_spq_left))
atomic_read      14601 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 			if (!atomic_read(&bp->eq_spq_left))
atomic_read      4166 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	c = atomic_read(v);
atomic_read      4194 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	c = atomic_read(v);
atomic_read      4238 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	cur_credit = atomic_read(&o->credit);
atomic_read      2144 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (unlikely(atomic_read(&bp->intr_sem) != 0))
atomic_read      9998 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (atomic_read(&bp->intr_sem) != 0)
atomic_read      1543 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	num_flows = atomic_read(&tc_info->flow_table.nelems);
atomic_read        90 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c 	while (atomic_read(&ulp->ref_count) != 0 && i < 10) {
atomic_read       548 drivers/net/ethernet/broadcom/cnic.c 	while ((atomic_read(&ulp_ops->ref_count) != 0) && (i < 20)) {
atomic_read       553 drivers/net/ethernet/broadcom/cnic.c 	if (atomic_read(&ulp_ops->ref_count) != 0)
atomic_read      3599 drivers/net/ethernet/broadcom/cnic.c 	if (atomic_read(&csk1->ref_count))
atomic_read      3654 drivers/net/ethernet/broadcom/cnic.c 	while (atomic_read(&csk->ref_count) != 1)
atomic_read      4230 drivers/net/ethernet/broadcom/cnic.c 	if (atomic_read(&cp->iscsi_conn) != 0)
atomic_read      4232 drivers/net/ethernet/broadcom/cnic.c 			    atomic_read(&cp->iscsi_conn));
atomic_read      5439 drivers/net/ethernet/broadcom/cnic.c 	while ((atomic_read(&dev->ref_count) != 0) && i < 10) {
atomic_read      5443 drivers/net/ethernet/broadcom/cnic.c 	if (atomic_read(&dev->ref_count) != 0)
atomic_read       562 drivers/net/ethernet/brocade/bna/bnad_debugfs.c 	if (atomic_read(&bna_debugfs_port_count) == 0) {
atomic_read      1432 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 	if (atomic_read(oct->adapter_refcount) > 1)
atomic_read      1501 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 	} while ((atomic_read(&ctx.status) == 0) && (count++ < timeout));
atomic_read      1503 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c 	ret = atomic_read(&ctx.status);
atomic_read       442 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 	} while ((!atomic_read(&status)) && (count++ < 100000));
atomic_read       444 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 	ret = atomic_read(&status);
atomic_read       765 drivers/net/ethernet/cavium/liquidio/lio_core.c 		if (atomic_read(&iq->instr_pending))
atomic_read       996 drivers/net/ethernet/cavium/liquidio/lio_core.c 	if (!(atomic_read(&oct->status) == OCT_DEV_IN_RESET))
atomic_read      1273 drivers/net/ethernet/cavium/liquidio/lio_core.c 				atomic_read(&oct->droq[idx]->pkts_pending);
atomic_read       269 drivers/net/ethernet/cavium/liquidio/lio_main.c 		if (atomic_read(&iq->instr_pending)) {
atomic_read       274 drivers/net/ethernet/cavium/liquidio/lio_main.c 				atomic_read(&iq->instr_pending);
atomic_read       911 drivers/net/ethernet/cavium/liquidio/lio_main.c 		if (atomic_read(oct_dev->adapter_refcount) == 1) {
atomic_read       989 drivers/net/ethernet/cavium/liquidio/lio_main.c 	switch (atomic_read(&oct->status)) {
atomic_read      1037 drivers/net/ethernet/cavium/liquidio/lio_main.c 			if (atomic_read(&iq->instr_pending)) {
atomic_read      1042 drivers/net/ethernet/cavium/liquidio/lio_main.c 					atomic_read(&iq->instr_pending);
atomic_read      1154 drivers/net/ethernet/cavium/liquidio/lio_main.c 		if (atomic_read(oct->adapter_fw_state) == FW_IS_PRELOADED)
atomic_read      1261 drivers/net/ethernet/cavium/liquidio/lio_main.c 	if (atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING)
atomic_read      1280 drivers/net/ethernet/cavium/liquidio/lio_main.c 	if (atomic_read(&lio->ifstate) & LIO_IFSTATE_REGISTERED)
atomic_read      2348 drivers/net/ethernet/cavium/liquidio/lio_main.c 	if (!(atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING) ||
atomic_read      3474 drivers/net/ethernet/cavium/liquidio/lio_main.c 		} else if (atomic_read(octeon_dev->adapter_fw_state) ==
atomic_read      3970 drivers/net/ethernet/cavium/liquidio/lio_main.c 	if (atomic_read(&oct->status) == OCT_DEV_RUNNING)
atomic_read      3978 drivers/net/ethernet/cavium/liquidio/lio_main.c 	if (atomic_read(&oct->status) != OCT_DEV_CORE_OK) {
atomic_read       129 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 		if (atomic_read(&iq->instr_pending)) {
atomic_read       134 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 			    atomic_read(&iq->instr_pending);
atomic_read       452 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	switch (atomic_read(&oct->status)) {
atomic_read       493 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 			if (atomic_read(&iq->instr_pending)) {
atomic_read       498 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 					atomic_read(&iq->instr_pending);
atomic_read       674 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	if (atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING)
atomic_read       692 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	if (atomic_read(&lio->ifstate) & LIO_IFSTATE_REGISTERED)
atomic_read      1433 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	if (!(atomic_read(&lio->ifstate) & LIO_IFSTATE_RUNNING) ||
atomic_read       132 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	atomic_set(&vf_rep->ifstate, (atomic_read(&vf_rep->ifstate) |
atomic_read       165 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	atomic_set(&vf_rep->ifstate, (atomic_read(&vf_rep->ifstate) &
atomic_read       317 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	if (!(atomic_read(&vf_rep->ifstate) & LIO_IFSTATE_RUNNING) ||
atomic_read       382 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	if (!(atomic_read(&vf_rep->ifstate) & LIO_IFSTATE_RUNNING) ||
atomic_read       631 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	s32 istate = (s32)atomic_read(state_ptr);
atomic_read       801 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	refcount = atomic_read(oct->adapter_refcount);
atomic_read       818 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	refcount = atomic_read(oct->adapter_refcount);
atomic_read      1204 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	if (atomic_read(&oct->status) >= OCT_DEV_RUNNING) {
atomic_read      1206 drivers/net/ethernet/cavium/liquidio/octeon_device.c 			atomic_read(&oct->status));
atomic_read      1441 drivers/net/ethernet/cavium/liquidio/octeon_device.c 		pkts_pend = (u32)atomic_read(&droq->pkts_pending);
atomic_read       736 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 		if (!atomic_read(&droq->pkts_pending))
atomic_read       752 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 	pkt_count = atomic_read(&droq->pkts_pending);
atomic_read       774 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 	if (atomic_read(&droq->pkts_pending))
atomic_read       800 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 				     (u32)(atomic_read(&droq->pkts_pending)));
atomic_read       506 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	return atomic_read(&lio->ifstate) & state_flag;
atomic_read       516 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	atomic_set(&lio->ifstate, (atomic_read(&lio->ifstate) | state_flag));
atomic_read       526 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	atomic_set(&lio->ifstate, (atomic_read(&lio->ifstate) & ~(state_flag)));
atomic_read       540 drivers/net/ethernet/cavium/liquidio/octeon_network.h 		pcount = atomic_read(
atomic_read       114 drivers/net/ethernet/cavium/liquidio/octeon_nic.h 	return ((u32)atomic_read(&oct->instr_queue[q_no]->instr_pending)
atomic_read       259 drivers/net/ethernet/cavium/liquidio/request_manager.c 			    atomic_read(&oct->instr_queue[i]->instr_pending);
atomic_read       278 drivers/net/ethernet/cavium/liquidio/request_manager.c 	if (atomic_read(&oct->status) == OCT_DEV_RUNNING) {
atomic_read       320 drivers/net/ethernet/cavium/liquidio/request_manager.c 	if (atomic_read(&iq->instr_pending) >= (s32)(iq->max_count - 1)) {
atomic_read       326 drivers/net/ethernet/cavium/liquidio/request_manager.c 	if (atomic_read(&iq->instr_pending) >= (s32)(iq->max_count - 2))
atomic_read       440 drivers/net/ethernet/cavium/liquidio/request_manager.c 	if (atomic_read(&oct->response_list
atomic_read       514 drivers/net/ethernet/cavium/liquidio/request_manager.c 	if (!atomic_read(&iq->instr_pending))
atomic_read       782 drivers/net/ethernet/cavium/liquidio/request_manager.c 	if (!atomic_read(&done_sc_list->pending_req_count))
atomic_read       231 drivers/net/ethernet/cavium/liquidio/response_manager.c 	if (atomic_read(&oct->response_list
atomic_read       937 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	    (atomic_read(&sq->free_cnt) >= MIN_SQ_DESC_PER_PKT_XMIT)) {
atomic_read      1300 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		if (atomic_read(&sq->free_cnt) > MIN_SQ_DESC_PER_PKT_XMIT) {
atomic_read      1558 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	if (subdesc_cnt > atomic_read(&sq->free_cnt))
atomic_read       658 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 	    t->atids_in_use + atomic_read(&t->tids_in_use) + MC5_MIN_TIDS <=
atomic_read       223 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	if (!atomic_read(&d->nfree))
atomic_read       228 drivers/net/ethernet/chelsio/cxgb3/l2t.c 		if (atomic_read(&e->refcnt) == 0)
atomic_read       231 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	for (e = &d->l2tab[1]; atomic_read(&e->refcnt); ++e) ;
atomic_read       267 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	if (atomic_read(&e->refcnt) == 0) {	/* hasn't been recycled */
atomic_read       338 drivers/net/ethernet/chelsio/cxgb3/l2t.c 			if (atomic_read(&e->refcnt) == 1)
atomic_read       421 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	if (atomic_read(&e->refcnt)) {
atomic_read       271 drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c 	seq_printf(seq, "Free clip entries : %d\n", atomic_read(&ctbl->nfree));
atomic_read      3135 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			   atomic_read(&t->conns_in_use));
atomic_read      3147 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 				   atomic_read(&t->tids_in_use),
atomic_read      3148 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 				   atomic_read(&t->hash_tids_in_use));
atomic_read      3156 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 				   atomic_read(&t->tids_in_use),
atomic_read      3157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 				   atomic_read(&t->hash_tids_in_use));
atomic_read      3163 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 				   atomic_read(&t->hash_tids_in_use));
atomic_read      3167 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			   atomic_read(&t->conns_in_use));
atomic_read      3172 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			   atomic_read(&t->tids_in_use));
atomic_read      3339 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.cipher_rqst));
atomic_read      3341 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.digest_rqst));
atomic_read      3343 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.aead_rqst));
atomic_read      3345 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.complete));
atomic_read      3347 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.error));
atomic_read      3349 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.fallback));
atomic_read      3351 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   atomic_read(&adap->chcr_stats.ipsec_cnt));
atomic_read       268 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	if (!atomic_read(&d->nfree))
atomic_read       273 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		if (atomic_read(&e->refcnt) == 0)
atomic_read       276 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	for (e = d->l2tab; atomic_read(&e->refcnt); ++e)
atomic_read       305 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		if (atomic_read(&e->refcnt) == 0) {
atomic_read       356 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	if (atomic_read(&e->refcnt) == 0) {  /* hasn't been recycled */
atomic_read       376 drivers/net/ethernet/chelsio/cxgb4/l2t.c 	if (atomic_read(&e->refcnt) == 0) {  /* hasn't been recycled */
atomic_read       448 drivers/net/ethernet/chelsio/cxgb4/l2t.c 			if (atomic_read(&e->refcnt) == 1)
atomic_read       548 drivers/net/ethernet/chelsio/cxgb4/l2t.c 			if (atomic_read(&e->refcnt))
atomic_read       596 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		if (!atomic_read(&e->refcnt)) {
atomic_read       725 drivers/net/ethernet/chelsio/cxgb4/l2t.c 			   l2e_state(e), atomic_read(&e->refcnt),
atomic_read       586 drivers/net/ethernet/emulex/benet/be_cmds.c 		if (atomic_read(&mcc_obj->q.used) == 0)
atomic_read       847 drivers/net/ethernet/emulex/benet/be_cmds.c 	if (atomic_read(&mccq->used) >= mccq->len)
atomic_read       819 drivers/net/ethernet/emulex/benet/be_main.c 	return atomic_read(&txo->q.used) + BE_MAX_TX_FRAG_COUNT >= txo->q.len;
atomic_read       824 drivers/net/ethernet/emulex/benet/be_main.c 	return atomic_read(&txo->q.used) < txo->q.len / 2;
atomic_read       829 drivers/net/ethernet/emulex/benet/be_main.c 	return atomic_read(&txo->q.used) > txo->pend_wrb_cnt;
atomic_read      1435 drivers/net/ethernet/emulex/benet/be_main.c 			 atomic_read(&txo->q.used), txo->q.id);
atomic_read      1450 drivers/net/ethernet/emulex/benet/be_main.c 			 atomic_read(&txo->cq.used));
atomic_read      2662 drivers/net/ethernet/emulex/benet/be_main.c 	} else if (atomic_read(&rxq->used) == 0) {
atomic_read      2817 drivers/net/ethernet/emulex/benet/be_main.c 	while (atomic_read(&rxq->used) > 0) {
atomic_read      2822 drivers/net/ethernet/emulex/benet/be_main.c 	BUG_ON(atomic_read(&rxq->used));
atomic_read      2911 drivers/net/ethernet/emulex/benet/be_main.c 		if (atomic_read(&txq->used)) {
atomic_read      2913 drivers/net/ethernet/emulex/benet/be_main.c 				 i, atomic_read(&txq->used));
atomic_read      2916 drivers/net/ethernet/emulex/benet/be_main.c 			index_adv(&end_idx, atomic_read(&txq->used) - 1,
atomic_read      2923 drivers/net/ethernet/emulex/benet/be_main.c 			BUG_ON(atomic_read(&txq->used));
atomic_read      3255 drivers/net/ethernet/emulex/benet/be_main.c 		if (atomic_read(&rxo->q.used) < RX_FRAGS_REFILL_WM &&
atomic_read      3582 drivers/net/ethernet/emulex/benet/be_main.c 				if (atomic_read(&q->used) == 0)
atomic_read        47 drivers/net/ethernet/google/gve/gve_tx.c 	WARN(atomic_read(&fifo->available) != fifo->size,
atomic_read        61 drivers/net/ethernet/google/gve/gve_tx.c 	return (atomic_read(&fifo->available) <= bytes) ? false : true;
atomic_read       218 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c 			if (atomic_read(&hdev->arq.count) >=
atomic_read       724 drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c 		HINIC_CMDQ_CTXT_BLOCK_INFO_SET(atomic_read(&wq->cons_idx), CI);
atomic_read       101 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	ci_start = atomic_read(&wq->cons_idx);
atomic_read       102 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	pi_start = atomic_read(&wq->prod_idx);
atomic_read       158 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	ci_start = atomic_read(&wq->cons_idx);
atomic_read       159 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	pi_start = atomic_read(&wq->prod_idx);
atomic_read       455 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	return atomic_read(&wq->delta) - 1;
atomic_read       468 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	return atomic_read(&wq->delta) - 1;
atomic_read       748 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c 	*prod_idx = MASKED_WQE_IDX(wq, atomic_read(&wq->prod_idx));
atomic_read       828 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c 	if ((atomic_read(&wq->delta) + num_wqebbs) > wq->q_depth)
atomic_read       831 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c 	curr_cons_idx = atomic_read(&wq->cons_idx);
atomic_read       258 drivers/net/ethernet/ibm/ehea/ehea_ethtool.c 		data[i++] = atomic_read(&port->port_res[k].swqe_avail);
atomic_read       858 drivers/net/ethernet/ibm/ehea/ehea_main.c 		     (atomic_read(&pr->swqe_avail) >= pr->swqe_refill_th))) {
atomic_read       861 drivers/net/ethernet/ibm/ehea/ehea_main.c 		    (atomic_read(&pr->swqe_avail) >= pr->swqe_refill_th))
atomic_read      2077 drivers/net/ethernet/ibm/ehea/ehea_main.c 	if (unlikely(atomic_read(&pr->swqe_avail) <= 1)) {
atomic_read      2495 drivers/net/ethernet/ibm/ehea/ehea_main.c 			 atomic_read(&pr->swqe_avail) >= swqe_max,
atomic_read      3351 drivers/net/ethernet/ibm/ehea/ehea_main.c 	if (atomic_read(&ehea_memory_hooks_registered) == 0)
atomic_read       209 drivers/net/ethernet/ibm/ibmveth.c 	u32 count = pool->size - atomic_read(&pool->available);
atomic_read       318 drivers/net/ethernet/ibm/ibmveth.c 		    (atomic_read(&pool->available) < pool->threshold))
atomic_read       245 drivers/net/ethernet/ibm/ibmvnic.c 	int count = pool->size - atomic_read(&pool->available);
atomic_read      4122 drivers/net/ethernet/ibm/ibmvnic.c 	if (atomic_read(&adapter->running_cap_crqs) == 0) {
atomic_read      4234 drivers/net/ethernet/ibm/ibmvnic.c 		   atomic_read(&adapter->running_cap_crqs));
atomic_read      4398 drivers/net/ethernet/ibm/ibmvnic.c 	if (atomic_read(&adapter->running_cap_crqs) == 0) {
atomic_read      4665 drivers/net/ethernet/ibm/ibmvnic.c 	if (atomic_read(&adapter->running_cap_crqs) != 0)
atomic_read      2356 drivers/net/ethernet/intel/e1000/e1000_main.c 	if (atomic_read(&adapter->tx_fifo_stall)) {
atomic_read      3048 drivers/net/ethernet/intel/e1000/e1000_main.c 	if (atomic_read(&adapter->tx_fifo_stall))
atomic_read      1063 drivers/net/ethernet/jme.c 	if (unlikely(atomic_read(&jme->link_changing) != 1))
atomic_read      1069 drivers/net/ethernet/jme.c 	i = atomic_read(&rxring->next_to_clean);
atomic_read      1201 drivers/net/ethernet/jme.c 		(atomic_read(&jme->link_changing) != 1)
atomic_read      1278 drivers/net/ethernet/jme.c 		while (atomic_read(&jme->link_changing) != 1)
atomic_read      1366 drivers/net/ethernet/jme.c 	while (atomic_read(&jme->rx_empty) > 0) {
atomic_read      1387 drivers/net/ethernet/jme.c 	if (unlikely(atomic_read(&jme->link_changing) != 1))
atomic_read      1397 drivers/net/ethernet/jme.c 	while (atomic_read(&jme->rx_empty) > 0) {
atomic_read      1412 drivers/net/ethernet/jme.c 	atomic_read(&txring->nr_free) >= (jme->tx_wake_threshold))) {
atomic_read      1433 drivers/net/ethernet/jme.c 	if (unlikely(atomic_read(&jme->link_changing) != 1))
atomic_read      1439 drivers/net/ethernet/jme.c 	max = jme->tx_ring_size - atomic_read(&txring->nr_free);
atomic_read      1442 drivers/net/ethernet/jme.c 	for (i = atomic_read(&txring->next_to_clean) ; cnt < max ; ) {
atomic_read      1955 drivers/net/ethernet/jme.c 	if (unlikely(atomic_read(&txring->nr_free) < nr_alloc))
atomic_read      2192 drivers/net/ethernet/jme.c 	int idx = atomic_read(&txring->next_to_clean);
atomic_read      2197 drivers/net/ethernet/jme.c 	if (unlikely(atomic_read(&txring->nr_free) < (MAX_SKB_FRAGS+2))) {
atomic_read      2201 drivers/net/ethernet/jme.c 		if (atomic_read(&txring->nr_free)
atomic_read      1145 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	if (unlikely(atomic_read(&ring->free_count) <= tx_num)) {
atomic_read      1171 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	if (unlikely(atomic_read(&ring->free_count) <= ring->thresh))
atomic_read      1450 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	    (atomic_read(&ring->free_count) > ring->thresh))
atomic_read      2704 drivers/net/ethernet/mellanox/mlx4/fw.c 	int num_tasks = atomic_read(&priv->opreq_count);
atomic_read      2050 drivers/net/ethernet/mellanox/mlx4/main.c 	if (atomic_read(&pf_loading)) {
atomic_read      1338 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	if (res->com.state == RES_QP_BUSY || atomic_read(&res->ref_count) ||
atomic_read      1341 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		       res->com.state, atomic_read(&res->ref_count));
atomic_read      1353 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	    atomic_read(&res->ref_count)) {
atomic_read      1357 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			 atomic_read(&res->ref_count));
atomic_read      1688 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		else if (atomic_read(&r->ref_count))
atomic_read      1728 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		else if (atomic_read(&r->ref_count))
atomic_read      4818 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		if (cq->com.owner == slave && !atomic_read(&cq->ref_count)) {
atomic_read      1774 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	wait_event(ctx->wait, atomic_read(&ctx->num_inflight) == 0);
atomic_read      4172 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	return atomic_read(&tc_ht->nelems);
atomic_read       359 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		if (atomic_read(&rep->rep_data[REP_ETH].state) != REP_LOADED)
atomic_read      2571 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	if (atomic_read(&rep->rep_data[rep_type].state) == REP_LOADED &&
atomic_read      1624 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		version += (u64)atomic_read(&iter->g->node.version);
atomic_read      1709 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (atomic_read(&ft->node.version) != ft_version) {
atomic_read      1779 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	version = atomic_read(&ft->node.version);
atomic_read      1810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	    version != atomic_read(&ft->node.version))
atomic_read      1596 drivers/net/ethernet/natsemi/ns83820.c 		dev->tx_done_idx, dev->tx_free_idx, atomic_read(&dev->nr_tx_skbs)
atomic_read      1605 drivers/net/ethernet/natsemi/ns83820.c 			atomic_read(&dev->nr_tx_skbs));
atomic_read       300 drivers/net/ethernet/netronome/nfp/flower/main.c 				atomic_read(replies) >= tot_repl,
atomic_read       504 drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c 	*data++ = atomic_read(&nn->ktls_no_space);
atomic_read       408 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c 		if (!atomic_read(&bar->refcnt))
atomic_read       805 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c 	if (WARN_ON(!atomic_read(&priv->refcnt)))
atomic_read       405 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c 	if (atomic_read(&area->refcount))
atomic_read       136 drivers/net/ethernet/qlogic/qed/qed_main.c 	if (atomic_read(&pdev->enable_cnt) == 1)
atomic_read       172 drivers/net/ethernet/qlogic/qed/qed_main.c 	if (atomic_read(&pdev->enable_cnt) == 1) {
atomic_read      2471 drivers/net/ethernet/qlogic/qla3xxx.c 	if (unlikely(atomic_read(&qdev->tx_count) < 2))
atomic_read       836 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 		if (atomic_read(&cmd->rsp_status) ==
atomic_read      2036 drivers/net/ethernet/sfc/ef10.c 		if (rc != -ENOENT || atomic_read(&efx->active_queues))
atomic_read      3952 drivers/net/ethernet/sfc/ef10.c 	WARN_ON(atomic_read(&efx->active_queues) < 0);
atomic_read      3981 drivers/net/ethernet/sfc/ef10.c 				   atomic_read(&efx->active_queues) == 0,
atomic_read      3983 drivers/net/ethernet/sfc/ef10.c 		pending = atomic_read(&efx->active_queues);
atomic_read      3110 drivers/net/ethernet/sfc/efx.c 	stats[GENERIC_STAT_rx_noskb_drops] = atomic_read(&efx->n_rx_noskb_drops);
atomic_read        49 drivers/net/ethernet/sfc/ethtool.c 	return atomic_read((atomic_t *) field);
atomic_read      2719 drivers/net/ethernet/sfc/falcon/efx.c 	stats[GENERIC_STAT_rx_noskb_drops] = atomic_read(&efx->n_rx_noskb_drops);
atomic_read        49 drivers/net/ethernet/sfc/falcon/ethtool.c 	return atomic_read((atomic_t *) field);
atomic_read       434 drivers/net/ethernet/sfc/falcon/farch.c 	WARN_ON(atomic_read(&tx_queue->flush_outstanding));
atomic_read       609 drivers/net/ethernet/sfc/falcon/farch.c 	return (atomic_read(&efx->active_queues) == 0 ||
atomic_read       610 drivers/net/ethernet/sfc/falcon/farch.c 		(atomic_read(&efx->rxq_flush_outstanding) < EF4_RX_FLUSH_COUNT
atomic_read       611 drivers/net/ethernet/sfc/falcon/farch.c 		 && atomic_read(&efx->rxq_flush_pending) > 0));
atomic_read       676 drivers/net/ethernet/sfc/falcon/farch.c 	while (timeout && atomic_read(&efx->active_queues) > 0) {
atomic_read       683 drivers/net/ethernet/sfc/falcon/farch.c 				if (atomic_read(&efx->rxq_flush_outstanding) >=
atomic_read       701 drivers/net/ethernet/sfc/falcon/farch.c 	if (atomic_read(&efx->active_queues) &&
atomic_read       704 drivers/net/ethernet/sfc/falcon/farch.c 			  "(rx %d+%d)\n", atomic_read(&efx->active_queues),
atomic_read       705 drivers/net/ethernet/sfc/falcon/farch.c 			  atomic_read(&efx->rxq_flush_outstanding),
atomic_read       706 drivers/net/ethernet/sfc/falcon/farch.c 			  atomic_read(&efx->rxq_flush_pending));
atomic_read      1147 drivers/net/ethernet/sfc/falcon/farch.c 	WARN_ON(atomic_read(&efx->active_queues) == 0);
atomic_read       362 drivers/net/ethernet/sfc/falcon/selftest.c 	if (atomic_read(&state->rx_bad) == 0) {
atomic_read       463 drivers/net/ethernet/sfc/falcon/selftest.c 	return atomic_read(&state->rx_good) == state->packet_count;
atomic_read       489 drivers/net/ethernet/sfc/falcon/selftest.c 	rx_good = atomic_read(&state->rx_good);
atomic_read       490 drivers/net/ethernet/sfc/falcon/selftest.c 	rx_bad = atomic_read(&state->rx_bad);
atomic_read       422 drivers/net/ethernet/sfc/farch.c 	WARN_ON(atomic_read(&tx_queue->flush_outstanding));
atomic_read       591 drivers/net/ethernet/sfc/farch.c 	return (atomic_read(&efx->active_queues) == 0 ||
atomic_read       592 drivers/net/ethernet/sfc/farch.c 		(atomic_read(&efx->rxq_flush_outstanding) < EFX_RX_FLUSH_COUNT
atomic_read       593 drivers/net/ethernet/sfc/farch.c 		 && atomic_read(&efx->rxq_flush_pending) > 0));
atomic_read       658 drivers/net/ethernet/sfc/farch.c 	while (timeout && atomic_read(&efx->active_queues) > 0) {
atomic_read       675 drivers/net/ethernet/sfc/farch.c 				if (atomic_read(&efx->rxq_flush_outstanding) >=
atomic_read       694 drivers/net/ethernet/sfc/farch.c 	if (atomic_read(&efx->active_queues) &&
atomic_read       697 drivers/net/ethernet/sfc/farch.c 			  "(rx %d+%d)\n", atomic_read(&efx->active_queues),
atomic_read       698 drivers/net/ethernet/sfc/farch.c 			  atomic_read(&efx->rxq_flush_outstanding),
atomic_read       699 drivers/net/ethernet/sfc/farch.c 			  atomic_read(&efx->rxq_flush_pending));
atomic_read      1135 drivers/net/ethernet/sfc/farch.c 	WARN_ON(atomic_read(&efx->active_queues) == 0);
atomic_read      1276 drivers/net/ethernet/sfc/mcdi_port.c 	if (rc && (rc != -ENOENT || atomic_read(&efx->active_queues)))
atomic_read       362 drivers/net/ethernet/sfc/selftest.c 	if (atomic_read(&state->rx_bad) == 0) {
atomic_read       463 drivers/net/ethernet/sfc/selftest.c 	return atomic_read(&state->rx_good) == state->packet_count;
atomic_read       489 drivers/net/ethernet/sfc/selftest.c 	rx_good = atomic_read(&state->rx_good);
atomic_read       490 drivers/net/ethernet/sfc/selftest.c 	rx_bad = atomic_read(&state->rx_bad);
atomic_read       663 drivers/net/ethernet/sfc/siena_sriov.c 		atomic_read(&vf->rxq_retry_count);
atomic_read       864 drivers/net/ethernet/silan/sc92031.c 	intr_mask = atomic_read(&priv->intr_mask);
atomic_read       897 drivers/net/ethernet/silan/sc92031.c 	intr_mask = atomic_read(&priv->intr_mask);
atomic_read      3834 drivers/net/ethernet/sun/cassini.c 	while (atomic_read(&cp->reset_task_pending_mtu) ||
atomic_read      3835 drivers/net/ethernet/sun/cassini.c 	       atomic_read(&cp->reset_task_pending_spare) ||
atomic_read      3836 drivers/net/ethernet/sun/cassini.c 	       atomic_read(&cp->reset_task_pending_all))
atomic_read      3840 drivers/net/ethernet/sun/cassini.c 	while (atomic_read(&cp->reset_task_pending))
atomic_read      3997 drivers/net/ethernet/sun/cassini.c 	int pending = atomic_read(&cp->reset_task_pending);
atomic_read      3999 drivers/net/ethernet/sun/cassini.c 	int pending_all = atomic_read(&cp->reset_task_pending_all);
atomic_read      4000 drivers/net/ethernet/sun/cassini.c 	int pending_spare = atomic_read(&cp->reset_task_pending_spare);
atomic_read      4001 drivers/net/ethernet/sun/cassini.c 	int pending_mtu = atomic_read(&cp->reset_task_pending_mtu);
atomic_read      4098 drivers/net/ethernet/sun/cassini.c 	if (atomic_read(&cp->reset_task_pending_all) ||
atomic_read      4099 drivers/net/ethernet/sun/cassini.c 	    atomic_read(&cp->reset_task_pending_spare) ||
atomic_read      4100 drivers/net/ethernet/sun/cassini.c 	    atomic_read(&cp->reset_task_pending_mtu))
atomic_read      4103 drivers/net/ethernet/sun/cassini.c 	if (atomic_read(&cp->reset_task_pending))
atomic_read       542 drivers/net/ethernet/ti/cpmac.c 	if (unlikely(atomic_read(&priv->reset_pending)))
atomic_read      1829 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		   atomic_read(&card->tx_timeout_task_counter) == 0);
atomic_read      2475 drivers/net/ethernet/toshiba/spider_net.c 		   atomic_read(&card->tx_timeout_task_counter) == 0);
atomic_read       875 drivers/net/hyperv/netvsc.c 	    atomic_read(&nvchan->queue_sends) < 1 &&
atomic_read      1006 drivers/net/hyperv/rndis_filter.c 		if (atomic_read(&nvchan->queue_sends) > 0)
atomic_read      1167 drivers/net/hyperv/rndis_filter.c 		   atomic_read(&nvdev->open_chn) == nvdev->num_chn);
atomic_read       383 drivers/net/plip/plip.c 	if (!(atomic_read (&nl->kill_timer))) {
atomic_read      3271 drivers/net/ppp/ppp_generic.c 	if (atomic_read(&ppp_unit_count) || atomic_read(&channel_count))
atomic_read       418 drivers/net/rionet.c 				state = atomic_read(&rdev->state);
atomic_read      1016 drivers/net/thunderbolt.c 	u16 frame_id = atomic_read(&net->frame_id);
atomic_read       263 drivers/net/usb/catc.c 		if (atomic_read(&catc->recq_sz)) {
atomic_read        48 drivers/net/usb/cdc_mbim.c 	dev_dbg(&dev->intf->dev, "%s() pmcount=%d, on=%d\n", __func__, atomic_read(&info->pmcount), on);
atomic_read      1347 drivers/net/usb/cdc_ncm.c 	if (!(hrtimer_active(&ctx->tx_timer) || atomic_read(&ctx->stop)))
atomic_read      1358 drivers/net/usb/cdc_ncm.c 	if (!atomic_read(&ctx->stop))
atomic_read       637 drivers/net/usb/qmi_wwan.c 		atomic_read(&info->pmcount), on);
atomic_read      1590 drivers/net/usb/r8152.c 	WARN_ON(atomic_read(&tp->rx_count));
atomic_read      2003 drivers/net/usb/r8152.c 	return atomic_read(&tp->rx_count) > RTL8152_MAX_RX;
atomic_read      2035 drivers/net/usb/r8152.c 	if (!agg_free && atomic_read(&tp->rx_count) < tp->rx_pending)
atomic_read       621 drivers/net/wimax/i2400m/driver.c 			if (atomic_read(&i2400m->bus_reset_retries)
atomic_read       201 drivers/net/wimax/i2400m/usb-rx.c 	do_autopm = atomic_read(&i2400mu->do_autopm);
atomic_read       351 drivers/net/wimax/i2400m/usb-rx.c 			 || (pending = atomic_read(&i2400mu->rx_pending_count)))
atomic_read        97 drivers/net/wimax/i2400m/usb-tx.c 	do_autopm = atomic_read(&i2400mu->do_autopm);
atomic_read       137 drivers/net/wireless/ath/ar5523/ar5523.c 			atomic_read(&ar->tx_nr_pending));
atomic_read       730 drivers/net/wireless/ath/ar5523/ar5523.c 	if (atomic_read(&ar->tx_nr_total) < AR5523_TX_DATA_RESTART_COUNT) {
atomic_read       776 drivers/net/wireless/ath/ar5523/ar5523.c 			   atomic_read(&ar->tx_nr_total),
atomic_read       777 drivers/net/wireless/ath/ar5523/ar5523.c 			   atomic_read(&ar->tx_nr_pending));
atomic_read       867 drivers/net/wireless/ath/ar5523/ar5523.c 			   atomic_read(&ar->tx_nr_pending));
atomic_read       911 drivers/net/wireless/ath/ar5523/ar5523.c 		   atomic_read(&ar->tx_nr_total),
atomic_read       912 drivers/net/wireless/ath/ar5523/ar5523.c 		   atomic_read(&ar->tx_nr_pending));
atomic_read       927 drivers/net/wireless/ath/ar5523/ar5523.c 	    !atomic_read(&ar->tx_nr_pending), AR5523_FLUSH_TIMEOUT))
atomic_read       929 drivers/net/wireless/ath/ar5523/ar5523.c 			   atomic_read(&ar->tx_nr_total),
atomic_read       930 drivers/net/wireless/ath/ar5523/ar5523.c 			   atomic_read(&ar->tx_nr_pending));
atomic_read      3984 drivers/net/wireless/ath/ath10k/htt_rx.c 	while (atomic_read(&htt->num_mpdus_ready)) {
atomic_read       611 drivers/net/wireless/ath/ath6kl/sdio.c 	return !atomic_read(&ar_sdio->irq_handling);
atomic_read       621 drivers/net/wireless/ath/ath6kl/sdio.c 	if (atomic_read(&ar_sdio->irq_handling)) {
atomic_read        83 drivers/net/wireless/ath/ath9k/htc_drv_init.c 	if (atomic_read(&priv->htc->tgt_ready) > 0) {
atomic_read       352 drivers/net/wireless/ath/ath9k/htc_drv_init.c 	if (atomic_read(&priv->wmi->mwrite_cnt))
atomic_read       506 drivers/net/wireless/ath/ath9k/htc_drv_init.c 	if (atomic_read(&priv->wmi->m_rmw_cnt))
atomic_read       881 drivers/net/wireless/ath/ath9k/mac.c 	if (atomic_read(&ah->intr_ref_cnt) != 0) {
atomic_read       883 drivers/net/wireless/ath/ath9k/mac.c 			atomic_read(&ah->intr_ref_cnt));
atomic_read       900 drivers/net/wireless/ath/ath9k/mac.c 			atomic_read(&ah->intr_ref_cnt));
atomic_read       226 drivers/net/wireless/ath/carl9170/debug.c 	    ar->fw.mem_blocks, atomic_read(&ar->mem_allocs));
atomic_read       229 drivers/net/wireless/ath/carl9170/debug.c 	    atomic_read(&ar->mem_free_blocks),
atomic_read       230 drivers/net/wireless/ath/carl9170/debug.c 	    (atomic_read(&ar->mem_free_blocks) * ar->fw.mem_block_size) / 1024,
atomic_read       677 drivers/net/wireless/ath/carl9170/debug.c 		atomic_read(&ar->pending_restarts));
atomic_read       773 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->tx_anch_urbs));
atomic_read       775 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->rx_anch_urbs));
atomic_read       777 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->rx_work_urbs));
atomic_read       779 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->rx_pool_urbs));
atomic_read       782 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->tx_total_queued));
atomic_read       784 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->tx_ampdu_scheduler));
atomic_read       787 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->tx_total_pending));
atomic_read       793 drivers/net/wireless/ath/carl9170/debug.c 		      atomic_read(&ar->tx_ampdu_upload));
atomic_read       245 drivers/net/wireless/ath/carl9170/main.c 	if (atomic_read(&ar->tx_total_queued))
atomic_read      1733 drivers/net/wireless/ath/carl9170/main.c 		if (atomic_read(&sta_info->pending_frames))
atomic_read      1747 drivers/net/wireless/ath/carl9170/main.c 	return !!atomic_read(&ar->tx_total_queued);
atomic_read        73 drivers/net/wireless/ath/carl9170/tx.c 		atomic_read(&ar->mem_free_blocks));
atomic_read       285 drivers/net/wireless/ath/carl9170/tx.c 	if (atomic_read(&ar->tx_total_queued))
atomic_read       289 drivers/net/wireless/ath/carl9170/tx.c 		if (!atomic_read(&ar->tx_ampdu_upload))
atomic_read       651 drivers/net/wireless/ath/carl9170/tx.c 	if (!atomic_read(&ar->tx_total_queued))
atomic_read      1139 drivers/net/wireless/ath/carl9170/tx.c 	if (atomic_read(&ar->tx_ampdu_upload))
atomic_read       321 drivers/net/wireless/ath/carl9170/usb.c 	while ((atomic_read(&ar->rx_anch_urbs) < AR9170_NUM_RX_URBS) &&
atomic_read       439 drivers/net/wireless/ath/carl9170/usb.c 		if (atomic_read(&ar->rx_anch_urbs) == 0) {
atomic_read       150 drivers/net/wireless/ath/wil6210/interrupt.c 	bool unmask_rx_htrsh = atomic_read(&wil->connected_vifs) > 0;
atomic_read       820 drivers/net/wireless/broadcom/b43/b43.h #define b43_status(wldev)		atomic_read(&(wldev)->__init_status)
atomic_read        62 drivers/net/wireless/broadcom/b43/leds.c 		turn_on = atomic_read(&led->state) != LED_OFF;
atomic_read       663 drivers/net/wireless/broadcom/b43legacy/b43legacy.h #define b43legacy_status(wldev)	atomic_read(&(wldev)->__init_status)
atomic_read       803 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 		WARN_ON(atomic_read(&sdiodev->freezer->freezing));
atomic_read       818 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 		   atomic_read(expect) == sdiodev->freezer->frozen_count);
atomic_read       836 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 	return atomic_read(&sdiodev->freezer->freezing);
atomic_read      1396 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c 	return atomic_read(&ifp->pend_8021x_cnt);
atomic_read       787 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	if ((force) || (atomic_read(&commonring->outstanding_tx) <
atomic_read      1383 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 		    ((qlen) && (atomic_read(&commonring->outstanding_tx) <
atomic_read      2511 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		if (!sdiodev->irq_en && !atomic_read(&bus->ipend)) {
atomic_read      2591 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	if (atomic_read(&bus->ipend) > 0) {
atomic_read      2676 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	if ((bus->clkstate == CLK_AVAIL) && !atomic_read(&bus->fcstate) &&
atomic_read      2697 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	} else if (atomic_read(&bus->intstatus) ||
atomic_read      2698 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		   atomic_read(&bus->ipend) > 0 ||
atomic_read      2699 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		   (!atomic_read(&bus->fcstate) &&
atomic_read       315 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	while (atomic_read(&wl->callbacks) > 0)
atomic_read      1440 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	callbacks = atomic_read(&wl->callbacks) - ret_val;
atomic_read      1448 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	SPINWAIT((atomic_read(&wl->callbacks) > callbacks), 100 * 1000);
atomic_read       851 drivers/net/wireless/intel/iwlegacy/debug.c 			      atomic_read(&il->queue_stop_count[cnt]));
atomic_read      1557 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c 		if (atomic_read(&sta_priv->pending_frames) > 0)
atomic_read      4903 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 					 atomic_read(&mvm->queue_sync_counter) == 0 ||
atomic_read       540 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	int pending = atomic_read(&rba->req_pending);
atomic_read       599 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 			pending = atomic_read(&rba->req_pending);
atomic_read      1461 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 			atomic_read(&trans_pcie->rba.req_pending) *
atomic_read       192 drivers/net/wireless/intersil/hostap/hostap_ap.c 	if (atomic_read(&sta->users) != 0)
atomic_read      1007 drivers/net/wireless/intersil/hostap/hostap_ap.c 		   sta->addr, atomic_read(&sta->users), sta->aid,
atomic_read       145 drivers/net/wireless/marvell/mwifiex/cmdevt.c 		atomic_read(&adapter->cmd_pending));
atomic_read       727 drivers/net/wireless/marvell/mwifiex/cmdevt.c 		command, atomic_read(&adapter->cmd_pending));
atomic_read      1121 drivers/net/wireless/marvell/mwifiex/cmdevt.c 	if (!adapter->cmd_sent && !atomic_read(&adapter->tx_hw_pending) &&
atomic_read      1128 drivers/net/wireless/marvell/mwifiex/cmdevt.c 			    atomic_read(&adapter->tx_hw_pending) ? "T" : "",
atomic_read       274 drivers/net/wireless/marvell/mwifiex/debugfs.c 		     atomic_read(&phist_data->num_samples));
atomic_read       288 drivers/net/wireless/marvell/mwifiex/debugfs.c 		value = atomic_read(&phist_data->rx_rate[i]);
atomic_read       296 drivers/net/wireless/marvell/mwifiex/debugfs.c 			value = atomic_read(&phist_data->rx_rate[i]);
atomic_read       304 drivers/net/wireless/marvell/mwifiex/debugfs.c 		value =  atomic_read(&phist_data->snr[i]);
atomic_read       309 drivers/net/wireless/marvell/mwifiex/debugfs.c 		value = atomic_read(&phist_data->noise_flr[i]);
atomic_read       315 drivers/net/wireless/marvell/mwifiex/debugfs.c 		value = atomic_read(&phist_data->sig_str[i]);
atomic_read       204 drivers/net/wireless/marvell/mwifiex/main.c 		    (atomic_read(&adapter->rx_pending) < LOW_RX_PENDING)) {
atomic_read       270 drivers/net/wireless/marvell/mwifiex/main.c 		if (atomic_read(&adapter->rx_pending) >= HIGH_RX_PENDING &&
atomic_read      1053 drivers/net/wireless/marvell/mwifiex/main.c 		if (atomic_read(&card->port[i].tx_data_urb_pending)) {
atomic_read      1116 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&cardp->tx_cmd_urb_pending));
atomic_read      1118 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&cardp->port[0].tx_data_urb_pending));
atomic_read      1120 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&cardp->port[1].tx_data_urb_pending));
atomic_read      1122 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&cardp->rx_cmd_urb_pending));
atomic_read      1124 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&cardp->rx_data_urb_pending));
atomic_read      1128 drivers/net/wireless/marvell/mwifiex/main.c 		     atomic_read(&adapter->tx_pending));
atomic_read      1130 drivers/net/wireless/marvell/mwifiex/main.c 		     atomic_read(&adapter->rx_pending));
atomic_read      1147 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&priv->wmm_tx_pending[0]));
atomic_read      1149 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&priv->wmm_tx_pending[1]));
atomic_read      1151 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&priv->wmm_tx_pending[2]));
atomic_read      1153 drivers/net/wireless/marvell/mwifiex/main.c 			     atomic_read(&priv->wmm_tx_pending[3]));
atomic_read      1425 drivers/net/wireless/marvell/mwifiex/main.c 	if (atomic_read(&adapter->rx_pending) ||
atomic_read      1426 drivers/net/wireless/marvell/mwifiex/main.c 	    atomic_read(&adapter->tx_pending) ||
atomic_read      1427 drivers/net/wireless/marvell/mwifiex/main.c 	    atomic_read(&adapter->cmd_pending)) {
atomic_read      1431 drivers/net/wireless/marvell/mwifiex/main.c 			    atomic_read(&adapter->rx_pending),
atomic_read      1432 drivers/net/wireless/marvell/mwifiex/main.c 			    atomic_read(&adapter->tx_pending),
atomic_read      1433 drivers/net/wireless/marvell/mwifiex/main.c 			    atomic_read(&adapter->cmd_pending));
atomic_read        72 drivers/net/wireless/marvell/mwifiex/tdls.c 		if (atomic_read(&priv->wmm.highest_queued_prio) <
atomic_read        59 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 			if ((atomic_read(&adapter->pending_bridged_pkts) <=
atomic_read       109 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 	if ((atomic_read(&adapter->pending_bridged_pkts) >=
atomic_read       309 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 			if ((atomic_read(&adapter->pending_bridged_pkts) >=
atomic_read       246 drivers/net/wireless/marvell/mwifiex/usb.c 		if (atomic_read(&adapter->rx_pending) <= HIGH_RX_PENDING) {
atomic_read       361 drivers/net/wireless/marvell/mwifiex/usb.c 	if (atomic_read(&card->rx_cmd_urb_pending) && card->rx_cmd.urb)
atomic_read       367 drivers/net/wireless/marvell/mwifiex/usb.c 	if (atomic_read(&card->rx_data_urb_pending))
atomic_read       568 drivers/net/wireless/marvell/mwifiex/usb.c 	if (atomic_read(&card->rx_cmd_urb_pending) && card->rx_cmd.urb)
atomic_read       571 drivers/net/wireless/marvell/mwifiex/usb.c 	if (atomic_read(&card->rx_data_urb_pending))
atomic_read       621 drivers/net/wireless/marvell/mwifiex/usb.c 	if (!atomic_read(&card->rx_data_urb_pending))
atomic_read       626 drivers/net/wireless/marvell/mwifiex/usb.c 	if (!atomic_read(&card->rx_cmd_urb_pending)) {
atomic_read       839 drivers/net/wireless/marvell/mwifiex/usb.c 	    atomic_read(&port->tx_data_urb_pending) ==
atomic_read      1080 drivers/net/wireless/marvell/mwifiex/usb.c 			if (atomic_read(&port->tx_data_urb_pending) >=
atomic_read      1141 drivers/net/wireless/marvell/mwifiex/usb.c 	if (atomic_read(&port->tx_data_urb_pending) >=
atomic_read      1193 drivers/net/wireless/marvell/mwifiex/usb.c 				if (atomic_read(&port->tx_data_urb_pending)
atomic_read      1539 drivers/net/wireless/marvell/mwifiex/usb.c 	    (!atomic_read(&card->rx_cmd_urb_pending)))
atomic_read       700 drivers/net/wireless/marvell/mwifiex/util.c 	if (atomic_read(&phist_data->num_samples) > MWIFIEX_HIST_MAX_SAMPLES)
atomic_read       485 drivers/net/wireless/marvell/mwifiex/wmm.c 		if (atomic_read(&priv->wmm.tx_pkts_queued))
atomic_read       662 drivers/net/wireless/marvell/mwifiex/wmm.c 		tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued);
atomic_read       706 drivers/net/wireless/marvell/mwifiex/wmm.c 		tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued);
atomic_read       887 drivers/net/wireless/marvell/mwifiex/wmm.c 	if (atomic_read(&priv->wmm.highest_queued_prio) <
atomic_read      1106 drivers/net/wireless/marvell/mwifiex/wmm.c 			    (atomic_read(&priv_tmp->wmm.tx_pkts_queued) == 0))
atomic_read      1115 drivers/net/wireless/marvell/mwifiex/wmm.c 			for (i = atomic_read(hqp); i >= LOW_PRIO_TID; --i) {
atomic_read      1135 drivers/net/wireless/marvell/mwifiex/wmm.c 			if (atomic_read(&priv_tmp->wmm.tx_pkts_queued) != 0) {
atomic_read      1152 drivers/net/wireless/marvell/mwifiex/wmm.c 	if (atomic_read(hqp) > i)
atomic_read      1530 drivers/net/wireless/marvell/mwifiex/wmm.c 			if (atomic_read(&adapter->tx_queued) >=
atomic_read      1560 drivers/net/wireless/marvell/mwl8k.c 	if (atomic_read(&priv->watchdog_event_pending))
atomic_read      1585 drivers/net/wireless/marvell/mwl8k.c 		if (atomic_read(&priv->watchdog_event_pending)) {
atomic_read        71 drivers/net/wireless/mediatek/mt7601u/debugfs.c 		   atomic_read(&dev->avg_ampdu_len));
atomic_read        26 drivers/net/wireless/quantenna/qtnfmac/util.h 	return atomic_read(&list->size);
atomic_read       221 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		seqno = atomic_read(&intf->seqno);
atomic_read        64 drivers/net/wireless/rsi/rsi_91x_coex.c 	} while (atomic_read(&coex_cb->coex_tx_thread.thread_done) == 0);
atomic_read       260 drivers/net/wireless/rsi/rsi_91x_main.c 	} while (atomic_read(&common->tx_thread.thread_done) == 0);
atomic_read        75 drivers/net/wireless/rsi/rsi_91x_sdio_ops.c 			if (atomic_read(&sdev->rx_thread.thread_done))
atomic_read        40 drivers/net/wireless/rsi/rsi_91x_usb_ops.c 			if (atomic_read(&dev->rx_thread.thread_done))
atomic_read        40 drivers/net/wireless/rsi/rsi_common.h 				(atomic_read(&event->event_condition) == 0));
atomic_read        43 drivers/net/wireless/rsi/rsi_common.h 				(atomic_read(&event->event_condition) == 0),
atomic_read       138 drivers/net/wireless/st/cw1200/bh.c 		(CW1200_BH_SUSPENDED == atomic_read(&priv->bh_suspend)),
atomic_read       153 drivers/net/wireless/st/cw1200/bh.c 		(CW1200_BH_RESUMED == atomic_read(&priv->bh_suspend)),
atomic_read       436 drivers/net/wireless/st/cw1200/bh.c 		    !atomic_read(&priv->recent_scan)) {
atomic_read       450 drivers/net/wireless/st/cw1200/bh.c 		    (atomic_read(&priv->bh_rx) == 0) &&
atomic_read       451 drivers/net/wireless/st/cw1200/bh.c 		    (atomic_read(&priv->bh_tx) == 0))
atomic_read       461 drivers/net/wireless/st/cw1200/bh.c 					0 : atomic_read(&priv->bh_suspend);
atomic_read       511 drivers/net/wireless/st/cw1200/bh.c 				   !atomic_read(&priv->recent_scan)) {
atomic_read       530 drivers/net/wireless/st/cw1200/bh.c 							  CW1200_BH_RESUME == atomic_read(&priv->bh_suspend));
atomic_read       232 drivers/net/wireless/st/cw1200/debug.c 		   atomic_read(&priv->bh_term) ? "terminated" : "alive");
atomic_read       234 drivers/net/wireless/st/cw1200/debug.c 		   atomic_read(&priv->bh_rx));
atomic_read       236 drivers/net/wireless/st/cw1200/debug.c 		   atomic_read(&priv->bh_tx));
atomic_read       260 drivers/net/wireless/st/cw1200/debug.c 		   atomic_read(&priv->tx_lock) ? "locked" : "unlocked");
atomic_read       261 drivers/net/wireless/st/cw1200/debug.c 	if (atomic_read(&priv->tx_lock))
atomic_read       263 drivers/net/wireless/st/cw1200/debug.c 			   atomic_read(&priv->tx_lock));
atomic_read       284 drivers/net/wireless/st/cw1200/debug.c 		   atomic_read(&priv->scan.in_progress) ? "active" : "idle");
atomic_read       149 drivers/net/wireless/st/cw1200/pm.c 	if (atomic_read(&priv->bh_rx)) {
atomic_read       271 drivers/net/wireless/st/cw1200/pm.c 	if (atomic_read(&priv->bh_rx)) {
atomic_read       159 drivers/net/wireless/st/cw1200/sta.c 		 atomic_read(&priv->tx_lock),
atomic_read       173 drivers/net/wireless/st/cw1200/sta.c 		if (!priv->vif->p2p && !atomic_read(&priv->tx_lock))
atomic_read      1243 drivers/net/wireless/st/cw1200/sta.c 	if (atomic_read(&priv->scan.in_progress)) {
atomic_read      1387 drivers/net/wireless/st/cw1200/sta.c 	if (atomic_read(&priv->scan.in_progress)) {
atomic_read      1186 drivers/net/wireless/st/cw1200/wsm.c 	BUG_ON(!atomic_read(&priv->tx_lock));
atomic_read       365 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (atomic_read(&intr->read_regs_enabled)) {
atomic_read       392 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	} else if (atomic_read(&intr->read_regs_enabled)) {
atomic_read       421 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (int_num == CR_INTERRUPT && atomic_read(&intr->read_regs_enabled))
atomic_read       464 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (hdr->id != USB_INT_ID_REGS && atomic_read(&intr->read_regs_enabled))
atomic_read      1027 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (!atomic_read(&tx->enabled)) {
atomic_read      1090 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (!atomic_read(&tx->enabled) || !tx->watchdog_enabled)
atomic_read      1488 drivers/net/xen-netback/netback.c 		!atomic_read(&queue->inflight_packets);
atomic_read      2075 drivers/net/xen-netfront.c 		data[i] = atomic_read((atomic_t *)(np + xennet_stats[i].offset));
atomic_read       846 drivers/ntb/test/ntb_perf.c 	return likely(atomic_read(&pthr->perf->tsync) > 0) ? 0 : -EINTR;
atomic_read       948 drivers/ntb/test/ntb_perf.c 		   (atomic_read(&pthr->dma_sync) == 0 ||
atomic_read       949 drivers/ntb/test/ntb_perf.c 		    atomic_read(&perf->tsync) < 0));
atomic_read       951 drivers/ntb/test/ntb_perf.c 	if (atomic_read(&perf->tsync) < 0)
atomic_read      1074 drivers/ntb/test/ntb_perf.c 				       atomic_read(&perf->tsync) <= 0);
atomic_read       175 drivers/ntb/test/ntb_pingpong.c 	count = atomic_read(&pp->count);
atomic_read       434 drivers/nvdimm/bus.c 			atomic_read(&nvdimm_bus->ioctl_active) == 0);
atomic_read       973 drivers/nvdimm/bus.c 	if (atomic_read(&nvdimm->busy))
atomic_read       342 drivers/nvdimm/dimm_devs.c 	return sprintf(buf, "%s\n", atomic_read(&nvdimm->busy)
atomic_read       534 drivers/nvdimm/security.c 		if (atomic_read(&nvdimm->busy)) {
atomic_read       542 drivers/nvdimm/security.c 		if (atomic_read(&nvdimm->busy)) {
atomic_read       455 drivers/nvme/host/fc.c 	if (atomic_read(&lport->act_rport_cnt) == 0)
atomic_read       855 drivers/nvme/host/fc.c 	if (atomic_read(&rport->act_ctrl_cnt) == 0)
atomic_read       769 drivers/nvme/target/fc.c 			    (!atomic_read(&queue->connected) ||
atomic_read      1762 drivers/nvme/target/fc.c 	sqtail = atomic_read(&q->sqtail) % q->sqsize;
atomic_read       170 drivers/oprofile/event_buffer.c 	wait_event_interruptible(buffer_wait, atomic_read(&buffer_ready));
atomic_read       176 drivers/oprofile/event_buffer.c 	if (!atomic_read(&buffer_ready))
atomic_read       184 drivers/oprofile/oprofilefs.c 	return oprofilefs_ulong_to_user(atomic_read(val), buf, count, offset);
atomic_read        88 drivers/pci/ats.c 	if (atomic_read(&dev->ats_ref_cnt))
atomic_read       420 drivers/pci/hotplug/cpci_hotplug_core.c 	    inserted, extracted, atomic_read(&extracting));
atomic_read       423 drivers/pci/hotplug/cpci_hotplug_core.c 	else if (!atomic_read(&extracting)) {
atomic_read       453 drivers/pci/hotplug/cpci_hotplug_core.c 		} while (atomic_read(&extracting) && !kthread_should_stop());
atomic_read       485 drivers/pci/hotplug/cpci_hotplug_core.c 			} while (atomic_read(&extracting) && !kthread_should_stop());
atomic_read       378 drivers/pci/hotplug/pciehp_ctrl.c 			   !atomic_read(&ctrl->pending_events) &&
atomic_read       412 drivers/pci/hotplug/pciehp_ctrl.c 			   !atomic_read(&ctrl->pending_events) &&
atomic_read       684 drivers/pci/hotplug/pciehp_hpc.c 		       atomic_read(&ctrl->pending_events))
atomic_read       313 drivers/pci/pci-sysfs.c 	return sprintf(buf, "%u\n", atomic_read(&pdev->enable_cnt));
atomic_read      1940 drivers/pci/pci.c 	dev_WARN_ONCE(&dev->dev, atomic_read(&dev->enable_cnt) <= 0,
atomic_read       897 drivers/pci/pcie/aer.c 	if (atomic_read(&dev->enable_cnt) == 0)
atomic_read        81 drivers/pci/switch/switchtec.c 	stuser->event_cnt = atomic_read(&stdev->event_cnt);
atomic_read       559 drivers/pci/switch/switchtec.c 	if (stuser->event_cnt != atomic_read(&stdev->event_cnt))
atomic_read       700 drivers/pci/switch/switchtec.c 	stuser->event_cnt = atomic_read(&stdev->event_cnt);
atomic_read      1325 drivers/pcmcia/ds.c 	if (atomic_read(&p_dev->socket->present) != 0)
atomic_read      1346 drivers/perf/arm-cci.c 		if (atomic_read(active_events) == 0)
atomic_read       305 drivers/phy/motorola/phy-cpcap-usb.c 	if (!atomic_read(&ddata->active))
atomic_read      3261 drivers/power/supply/ab8500_charger.c 	if (atomic_read(&di->current_stepping_sessions))
atomic_read       549 drivers/power/supply/cpcap-battery.c 	if (!atomic_read(&ddata->active))
atomic_read       481 drivers/power/supply/cpcap-charger.c 	if (!atomic_read(&ddata->active))
atomic_read       421 drivers/power/supply/power_supply_core.c 	if (atomic_read(&psy->use_cnt) >= 0 &&
atomic_read       776 drivers/power/supply/power_supply_core.c 	if (atomic_read(&psy->use_cnt) <= 0) {
atomic_read       790 drivers/power/supply/power_supply_core.c 	if (atomic_read(&psy->use_cnt) <= 0 || !psy->desc->set_property)
atomic_read       800 drivers/power/supply/power_supply_core.c 	if (atomic_read(&psy->use_cnt) <= 0 ||
atomic_read       810 drivers/power/supply/power_supply_core.c 	if (atomic_read(&psy->use_cnt) <= 0 ||
atomic_read      1890 drivers/rapidio/devices/rio_mport_cdev.c 	if (atomic_read(&chdev->active) == 0)
atomic_read      2069 drivers/rapidio/devices/rio_mport_cdev.c 	if (atomic_read(&md->active) == 0)
atomic_read      2043 drivers/rapidio/rio_cm.c 			if (atomic_read(&rdev->state) != RIO_DEVICE_SHUTDOWN)
atomic_read        12 drivers/ras/debugfs.c 	return atomic_read(&trace_count);
atomic_read        18 drivers/ras/debugfs.c 	return atomic_read(&trace_count);
atomic_read       618 drivers/remoteproc/stm32_rproc.c 	if (atomic_read(&rproc->power) > 0)
atomic_read       306 drivers/reset/core.c 		if (WARN_ON(atomic_read(&rstc->deassert_count) != 0))
atomic_read       353 drivers/reset/core.c 		if (WARN_ON(atomic_read(&rstc->triggered_count) != 0))
atomic_read       356 drivers/reset/core.c 		if (WARN_ON(atomic_read(&rstc->deassert_count) == 0))
atomic_read       412 drivers/reset/core.c 		if (WARN_ON(atomic_read(&rstc->triggered_count) != 0))
atomic_read      3662 drivers/s390/block/dasd.c 		open_count = atomic_read(&device->block->open_count);
atomic_read      2266 drivers/s390/block/dasd_3990_erp.c 	if (atomic_read(&device->path[pos].error_count) >=
atomic_read       622 drivers/s390/block/dasd_devmap.c 	wait_event(dasd_delete_wq, atomic_read(&device->ref_count) == 0);
atomic_read       502 drivers/s390/block/dasd_ioctl.c 	dasd_info->open_count = atomic_read(&block->open_count);
atomic_read       355 drivers/s390/block/dcssblk.c 	if (atomic_read(&dev_info->use_count)) {
atomic_read       462 drivers/s390/block/dcssblk.c 		if (atomic_read(&dev_info->use_count) == 0) {
atomic_read       777 drivers/s390/block/dcssblk.c 	if (atomic_read(&dev_info->use_count) != 0) {
atomic_read       213 drivers/s390/char/monreader.c 	if (!atomic_read(&monpriv->read_ready))
atomic_read       313 drivers/s390/char/monreader.c 		   atomic_read(&monpriv->iucv_connected) ||
atomic_read       314 drivers/s390/char/monreader.c 		   atomic_read(&monpriv->iucv_severed));
atomic_read       315 drivers/s390/char/monreader.c 	if (atomic_read(&monpriv->iucv_severed)) {
atomic_read       382 drivers/s390/char/monreader.c 					atomic_read(&monpriv->read_ready) ||
atomic_read       383 drivers/s390/char/monreader.c 					atomic_read(&monpriv->iucv_severed));
atomic_read       386 drivers/s390/char/monreader.c 		if (unlikely(atomic_read(&monpriv->iucv_severed)))
atomic_read       437 drivers/s390/char/monreader.c 	if (unlikely(atomic_read(&monpriv->iucv_severed)))
atomic_read       439 drivers/s390/char/monreader.c 	if (atomic_read(&monpriv->read_ready))
atomic_read       506 drivers/s390/char/monreader.c 		   atomic_read(&monpriv->iucv_connected) ||
atomic_read       507 drivers/s390/char/monreader.c 		   atomic_read(&monpriv->iucv_severed));
atomic_read       508 drivers/s390/char/monreader.c 	if (atomic_read(&monpriv->iucv_severed))
atomic_read      1015 drivers/s390/char/raw3270.c 	wait_event(raw3270_wait_queue, atomic_read(&view->ref_count) == 0);
atomic_read       417 drivers/s390/char/vmlogrdr.c 	if (atomic_read(&priv->receive_ready)) {
atomic_read       486 drivers/s390/char/vmlogrdr.c 					atomic_read(&priv->receive_ready));
atomic_read       494 drivers/s390/cio/ccwgroup.c 	if (atomic_read(&gdev->onoff))
atomic_read        71 drivers/s390/cio/crw.c 					  atomic_read(&crw_nr_req) > 0);
atomic_read       146 drivers/s390/cio/crw.c 	wait_event(crw_handler_wait_q, atomic_read(&crw_nr_req) == 0);
atomic_read      1328 drivers/s390/cio/css.c 				       atomic_read(&css_eval_scheduled) == 0);
atomic_read       170 drivers/s390/cio/device.c 				atomic_read(&ccw_device_init_count) == 0);
atomic_read      1808 drivers/s390/cio/device.c 	if (atomic_read(&cdev->private->onoff))
atomic_read       125 drivers/s390/cio/qdio_debug.c 		   atomic_read(&q->nr_buf_used), q->first_to_check);
atomic_read       511 drivers/s390/cio/qdio_main.c 	count = min(atomic_read(&q->nr_buf_used), QDIO_MAX_BUFFERS_MASK);
atomic_read       568 drivers/s390/cio/qdio_main.c 	if (!atomic_read(&q->nr_buf_used))
atomic_read       722 drivers/s390/cio/qdio_main.c 	count = atomic_read(&q->nr_buf_used);
atomic_read       766 drivers/s390/cio/qdio_main.c 	return atomic_read(&q->nr_buf_used) == 0;
atomic_read       832 drivers/s390/cio/qdio_main.c 	WARN_ON_ONCE(atomic_read(&q->nr_buf_used) < 0);
atomic_read       138 drivers/s390/cio/qdio_thinint.c 	if (!atomic_read(&q_indicators[TIQDIO_SHARED_IND].count))
atomic_read        92 drivers/s390/cio/vfio_ccw_ops.c 	return sprintf(buf, "%d\n", atomic_read(&private->avail));
atomic_read       379 drivers/s390/crypto/vfio_ap_ops.c 		       atomic_read(&matrix_dev->available_instances));
atomic_read        80 drivers/s390/crypto/zcrypt_api.c 	if (atomic_read(&zcrypt_rescan_req)) {
atomic_read       605 drivers/s390/crypto/zcrypt_api.c 	weight += atomic_read(&zc->load);
atomic_read       606 drivers/s390/crypto/zcrypt_api.c 	pref_weight += atomic_read(&pref_zc->load);
atomic_read       620 drivers/s390/crypto/zcrypt_api.c 	weight += atomic_read(&zq->load);
atomic_read       621 drivers/s390/crypto/zcrypt_api.c 	pref_weight += atomic_read(&pref_zq->load);
atomic_read      1432 drivers/s390/crypto/zcrypt_api.c 		return put_user(atomic_read(&zcrypt_open_count),
atomic_read        89 drivers/s390/crypto/zcrypt_card.c 	return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&zc->load));
atomic_read        81 drivers/s390/crypto/zcrypt_queue.c 	return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&zq->load));
atomic_read       126 drivers/s390/net/fsm.c 	int st = atomic_read(&fi->state);
atomic_read       147 drivers/s390/net/fsm.h 	int state = atomic_read(&fi->state);
atomic_read       215 drivers/s390/net/fsm.h 	return atomic_read(&fi->state);
atomic_read       553 drivers/s390/net/qeth_core.h 	return atomic_read(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q;
atomic_read       558 drivers/s390/net/qeth_core.h 	return atomic_read(&queue->used_buffers) == 0;
atomic_read       407 drivers/s390/net/qeth_core_main.c 			    atomic_read(&c->state) ==
atomic_read       428 drivers/s390/net/qeth_core_main.c 	if (forced_cleanup && (atomic_read(&(q->bufs[bidx]->state)) ==
atomic_read       455 drivers/s390/net/qeth_core_main.c 		WARN_ON_ONCE(atomic_read(&buffer->state) !=
atomic_read      1098 drivers/s390/net/qeth_core_main.c 	WARN_ON_ONCE(atomic_read(&buf->state) == QETH_QDIO_BUF_IN_CQ);
atomic_read      1100 drivers/s390/net/qeth_core_main.c 	if (atomic_read(&buf->state) == QETH_QDIO_BUF_PENDING)
atomic_read      1241 drivers/s390/net/qeth_core_main.c 	if (atomic_read(&card->qdio.state) != QETH_QDIO_UNINITIALIZED)
atomic_read      3227 drivers/s390/net/qeth_core_main.c 	if ((atomic_read(&buffer->state) == QETH_QDIO_BUF_EMPTY) &&
atomic_read      3245 drivers/s390/net/qeth_core_main.c 		if (atomic_read(&queue->used_buffers)
atomic_read      3264 drivers/s390/net/qeth_core_main.c 		if (atomic_read(&queue->used_buffers)
atomic_read      3298 drivers/s390/net/qeth_core_main.c 			if ((atomic_read(&queue->used_buffers) >=
atomic_read      3301 drivers/s390/net/qeth_core_main.c 			    !atomic_read(&queue->set_pci_flags_count)) {
atomic_read      3308 drivers/s390/net/qeth_core_main.c 			if (!atomic_read(&queue->set_pci_flags_count)) {
atomic_read      3324 drivers/s390/net/qeth_core_main.c 	if (atomic_read(&queue->set_pci_flags_count))
atomic_read      3368 drivers/s390/net/qeth_core_main.c 	if ((atomic_read(&queue->used_buffers) <= QETH_LOW_WATERMARK_PACK) ||
atomic_read      3369 drivers/s390/net/qeth_core_main.c 	    !atomic_read(&queue->set_pci_flags_count)) {
atomic_read      3383 drivers/s390/net/qeth_core_main.c 			    !atomic_read(&queue->set_pci_flags_count))
atomic_read      3844 drivers/s390/net/qeth_core_main.c 	if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY)
atomic_read      3854 drivers/s390/net/qeth_core_main.c 		if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY)
atomic_read      3909 drivers/s390/net/qeth_core_main.c 	if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) {
atomic_read      3932 drivers/s390/net/qeth_core_main.c 			if (atomic_read(&buffer->state) !=
atomic_read      3967 drivers/s390/net/qeth_core_main.c 	else if (!atomic_read(&queue->set_pci_flags_count))
atomic_read      3983 drivers/s390/net/qeth_core_main.c 		if (!tmp && !atomic_read(&queue->set_pci_flags_count))
atomic_read      5087 drivers/s390/net/qeth_core_main.c 	     (!atomic_read(&card->force_alloc_skb))) ||
atomic_read       277 drivers/s390/scsi/zfcp_aux.c 			if (atomic_read(&adapter->stat_miss) >=
atomic_read       250 drivers/s390/scsi/zfcp_ccw.c 		if (atomic_read(&adapter->status) &
atomic_read       264 drivers/s390/scsi/zfcp_ccw.c 		if (atomic_read(&adapter->status) &
atomic_read       254 drivers/s390/scsi/zfcp_dbf.c 	rec->adapter_status = atomic_read(&adapter->status);
atomic_read       256 drivers/s390/scsi/zfcp_dbf.c 		rec->port_status = atomic_read(&port->status);
atomic_read       261 drivers/s390/scsi/zfcp_dbf.c 		rec->lun_status = atomic_read(&sdev_to_zfcp(sdev)->status);
atomic_read       364 drivers/s390/scsi/zfcp_dbf.c 			atomic_read(&sdev_to_zfcp(erp->sdev)->erp_counter);
atomic_read       366 drivers/s390/scsi/zfcp_dbf.c 		rec->u.run.rec_count = atomic_read(&erp->port->erp_counter);
atomic_read       368 drivers/s390/scsi/zfcp_dbf.c 		rec->u.run.rec_count = atomic_read(&erp->adapter->erp_counter);
atomic_read       339 drivers/s390/scsi/zfcp_def.h 	return atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_MB_ACT;
atomic_read        87 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_ERP_INUSE)
atomic_read        95 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_INUSE)
atomic_read       110 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_INUSE)
atomic_read       130 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&zsdev->status) & ZFCP_STATUS_COMMON_ERP_FAILED)
atomic_read       134 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED)
atomic_read       138 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&port->status) &
atomic_read       147 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&adapter->status) &
atomic_read       172 drivers/s390/scsi/zfcp_erp.c 		l_status = atomic_read(&zfcp_sdev->status);
atomic_read       175 drivers/s390/scsi/zfcp_erp.c 		p_status = atomic_read(&port->status);
atomic_read       183 drivers/s390/scsi/zfcp_erp.c 		p_status = atomic_read(&port->status);
atomic_read       188 drivers/s390/scsi/zfcp_erp.c 		p_status = atomic_read(&port->status);
atomic_read       191 drivers/s390/scsi/zfcp_erp.c 		a_status = atomic_read(&adapter->status);
atomic_read       201 drivers/s390/scsi/zfcp_erp.c 		a_status = atomic_read(&adapter->status);
atomic_read       238 drivers/s390/scsi/zfcp_erp.c 		if (!(atomic_read(&zfcp_sdev->status) &
atomic_read       252 drivers/s390/scsi/zfcp_erp.c 		if (!(atomic_read(&port->status) & ZFCP_STATUS_COMMON_RUNNING))
atomic_read       263 drivers/s390/scsi/zfcp_erp.c 		if (!(atomic_read(&adapter->status) &
atomic_read       529 drivers/s390/scsi/zfcp_erp.c 	return (atomic_read(status) ^ mask) & mask;
atomic_read       757 drivers/s390/scsi/zfcp_erp.c 		if (!(atomic_read(&adapter->status) &
atomic_read       768 drivers/s390/scsi/zfcp_erp.c 	if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_XCONFIG_OK))
atomic_read       870 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_OPEN) {
atomic_read       903 drivers/s390/scsi/zfcp_erp.c 	int status = atomic_read(&port->status);
atomic_read       973 drivers/s390/scsi/zfcp_erp.c 	int p_status = atomic_read(&port->status);
atomic_read      1014 drivers/s390/scsi/zfcp_erp.c 	int p_status = atomic_read(&port->status);
atomic_read      1086 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_OPEN)
atomic_read      1091 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_OPEN)
atomic_read      1098 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_OPEN)
atomic_read      1122 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&zfcp_sdev->erp_counter) > ZFCP_MAX_ERPS) {
atomic_read      1140 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_ERP_FAILED) {
atomic_read      1157 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_NOESC) {
atomic_read      1162 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&port->erp_counter) > ZFCP_MAX_ERPS) {
atomic_read      1178 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED) {
atomic_read      1196 drivers/s390/scsi/zfcp_erp.c 		if (atomic_read(&adapter->erp_counter) > ZFCP_MAX_ERPS) {
atomic_read      1212 drivers/s390/scsi/zfcp_erp.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_FAILED) {
atomic_read      1246 drivers/s390/scsi/zfcp_erp.c 	int status = atomic_read(target_status);
atomic_read      1349 drivers/s390/scsi/zfcp_erp.c 	port_status = atomic_read(&port->status);
atomic_read      1371 drivers/s390/scsi/zfcp_erp.c 		lun_status = atomic_read(&zsdev->status);
atomic_read      1611 drivers/s390/scsi/zfcp_erp.c 		   !(atomic_read(&adapter->status) &
atomic_read       178 drivers/s390/scsi/zfcp_fc.c 	if ((atomic_read(&wka_port->refcount) != 0) ||
atomic_read       518 drivers/s390/scsi/zfcp_fc.c 	    !(atomic_read(&port->status) & ZFCP_STATUS_COMMON_OPEN)) {
atomic_read       581 drivers/s390/scsi/zfcp_fc.c 	if (atomic_read(&port->status) & ZFCP_STATUS_PORT_LINK_TEST)
atomic_read       699 drivers/s390/scsi/zfcp_fc.c 	if (!(atomic_read(&port->status) & ZFCP_STATUS_COMMON_NOESC))
atomic_read      1057 drivers/s390/scsi/zfcp_fc.c 	if (!(atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_OPEN))
atomic_read       126 drivers/s390/scsi/zfcp_fsf.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED)
atomic_read       440 drivers/s390/scsi/zfcp_fsf.c 	BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP);
atomic_read       762 drivers/s390/scsi/zfcp_fsf.c 	req->qdio_req.qdio_outb_usage = atomic_read(&qdio->req_q_free);
atomic_read       929 drivers/s390/scsi/zfcp_fsf.c 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
atomic_read      2282 drivers/s390/scsi/zfcp_fsf.c 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
atomic_read      2287 drivers/s390/scsi/zfcp_fsf.c 	if (atomic_read(&qdio->req_q_free) <= 0) {
atomic_read      2393 drivers/s390/scsi/zfcp_fsf.c 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
atomic_read        57 drivers/s390/scsi/zfcp_qdio.c 	used = QDIO_MAX_BUFFERS_PER_Q - atomic_read(&qdio->req_q_free);
atomic_read       210 drivers/s390/scsi/zfcp_qdio.c 	if (atomic_read(&qdio->req_q_free) ||
atomic_read       211 drivers/s390/scsi/zfcp_qdio.c 	    !(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP))
atomic_read       233 drivers/s390/scsi/zfcp_qdio.c 	if (!(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP))
atomic_read       347 drivers/s390/scsi/zfcp_qdio.c 	if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP))
atomic_read       360 drivers/s390/scsi/zfcp_qdio.c 	count = atomic_read(&qdio->req_q_free);
atomic_read       384 drivers/s390/scsi/zfcp_qdio.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)
atomic_read       497 drivers/s390/scsi/zfcp_qdio.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_SIOSL_ISSUED)
atomic_read       114 drivers/s390/scsi/zfcp_qdio.h 	int count = min(atomic_read(&qdio->req_q_free),
atomic_read       209 drivers/s390/scsi/zfcp_qdio.h 	int count = min(atomic_read(&qdio->req_q_free), max_sbals);
atomic_read        85 drivers/s390/scsi/zfcp_scsi.c 	status = atomic_read(&zfcp_sdev->status);
atomic_read        87 drivers/s390/scsi/zfcp_scsi.c 		     !(atomic_read(&zfcp_sdev->port->status) &
atomic_read       204 drivers/s390/scsi/zfcp_scsi.c 		if (!(atomic_read(&adapter->status) &
atomic_read       311 drivers/s390/scsi/zfcp_scsi.c 		if (!(atomic_read(&adapter->status) &
atomic_read       651 drivers/s390/scsi/zfcp_scsi.c 	int status = atomic_read(&adapter->status);
atomic_read       808 drivers/s390/scsi/zfcp_scsi.c 	data_div = atomic_read(&adapter->status) &
atomic_read        60 drivers/s390/scsi/zfcp_sysfs.c ZFCP_DEFINE_A_ATTR(status, "0x%08x\n", atomic_read(&adapter->status));
atomic_read        69 drivers/s390/scsi/zfcp_sysfs.c ZFCP_DEFINE_A_ATTR(in_recovery, "%d\n", (atomic_read(&adapter->status) &
atomic_read        73 drivers/s390/scsi/zfcp_sysfs.c 		 atomic_read(&port->status));
atomic_read        75 drivers/s390/scsi/zfcp_sysfs.c 		 (atomic_read(&port->status) &
atomic_read        96 drivers/s390/scsi/zfcp_sysfs.c 	if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED)
atomic_read       132 drivers/s390/scsi/zfcp_sysfs.c 		status = atomic_read(&sdev_to_zfcp(sdev)->status);
atomic_read       177 drivers/s390/scsi/zfcp_sysfs.c 	if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_FAILED)
atomic_read       247 drivers/s390/scsi/zfcp_sysfs.c 	return atomic_read(&port->units) == -1;
atomic_read       258 drivers/s390/scsi/zfcp_sysfs.c 	if (atomic_read(&port->units) > 0)
atomic_read       506 drivers/s390/scsi/zfcp_sysfs.c 		      (atomic_read(&zfcp_sdev->status) &
atomic_read       514 drivers/s390/scsi/zfcp_sysfs.c 	unsigned int status = atomic_read(&sdev_to_zfcp(sdev)->status);
atomic_read       542 drivers/s390/scsi/zfcp_sysfs.c 		      (atomic_read(&zfcp_sdev->status) &
atomic_read       546 drivers/s390/scsi/zfcp_sysfs.c 		      atomic_read(&zfcp_sdev->status));
atomic_read       654 drivers/s390/scsi/zfcp_sysfs.c 	return sprintf(buf, "%d %llu\n", atomic_read(&qdio->req_q_full),
atomic_read       223 drivers/s390/scsi/zfcp_unit.c 		status = atomic_read(&zfcp_sdev->status);
atomic_read       395 drivers/scsi/aacraid/commsup.c 				qid, atomic_read(&q->numpending));
atomic_read       277 drivers/scsi/arcmsr/arcmsr_attr.c 			atomic_read(&acb->ccboutstandingcount));
atomic_read      1402 drivers/scsi/arcmsr/arcmsr_hba.c 				, atomic_read(&acb->ccboutstandingcount));
atomic_read      1482 drivers/scsi/arcmsr/arcmsr_hba.c 		residual = atomic_read(&acb->ccboutstandingcount);
atomic_read      1615 drivers/scsi/arcmsr/arcmsr_hba.c 		if (!atomic_read(&acb->ccboutstandingcount))
atomic_read      1621 drivers/scsi/arcmsr/arcmsr_hba.c 	if (atomic_read(&acb->ccboutstandingcount)) {
atomic_read      3357 drivers/scsi/arcmsr/arcmsr_hba.c 				, atomic_read(&acb->ccboutstandingcount));
atomic_read      3426 drivers/scsi/arcmsr/arcmsr_hba.c 				, atomic_read(&acb->ccboutstandingcount));
atomic_read      3488 drivers/scsi/arcmsr/arcmsr_hba.c 				, atomic_read(&acb->ccboutstandingcount));
atomic_read      3565 drivers/scsi/arcmsr/arcmsr_hba.c 				, atomic_read(&acb->ccboutstandingcount));
atomic_read      3633 drivers/scsi/arcmsr/arcmsr_hba.c 				, atomic_read(&acb->ccboutstandingcount));
atomic_read      3968 drivers/scsi/arcmsr/arcmsr_hba.c 	if (unlikely(atomic_read(&acb->rq_map_token) == 0) ||
atomic_read      3975 drivers/scsi/arcmsr/arcmsr_hba.c 		if (atomic_read(&acb->ante_token_value) ==
atomic_read      3976 drivers/scsi/arcmsr/arcmsr_hba.c 			atomic_read(&acb->rq_map_token)) {
atomic_read      3980 drivers/scsi/arcmsr/arcmsr_hba.c 			atomic_read(&acb->rq_map_token));
atomic_read      4341 drivers/scsi/arcmsr/arcmsr_hba.c 	if (atomic_read(&acb->ccboutstandingcount) != 0) {
atomic_read      4454 drivers/scsi/arcmsr/arcmsr_hba.c 	if (!atomic_read(&acb->ccboutstandingcount)) {
atomic_read       500 drivers/scsi/bfa/bfad_debugfs.c 	if (atomic_read(&bfa_debugfs_port_count) == 0) {
atomic_read       424 drivers/scsi/bnx2fc/bnx2fc_io.c 	free_sqes = atomic_read(&tgt->free_sqes);
atomic_read       483 drivers/scsi/bnx2fc/bnx2fc_io.c 	free_sqes = atomic_read(&tgt->free_sqes);
atomic_read       160 drivers/scsi/bnx2i/bnx2i_hwi.c 		num_active_cmds = atomic_read(&ep->num_active_cmds);
atomic_read      2045 drivers/scsi/bnx2i/bnx2i_hwi.c 			if (!atomic_read(&bnx2i_conn->ep->num_active_cmds))
atomic_read      1234 drivers/scsi/bnx2i/bnx2i_iscsi.c 	if (atomic_read(&bnx2i_conn->ep->num_active_cmds) + 1  >
atomic_read      1490 drivers/scsi/bnx2i/bnx2i_iscsi.c 	if (atomic_read(&bnx2i_conn->work_cnt)) {
atomic_read       706 drivers/scsi/cxlflash/main.c 		while (atomic_read(&afu->cmds_active))
atomic_read      1705 drivers/scsi/cxlflash/superpipe.c 			    ((atomic_read(&cfg->recovery_threads) > 1) ||
atomic_read        83 drivers/scsi/esas2r/esas2r_int.c 	if (atomic_read(&a->disable_cnt) == 0)
atomic_read       127 drivers/scsi/esas2r/esas2r_int.c 	if (likely(atomic_read(&a->disable_cnt) == 0))
atomic_read       162 drivers/scsi/esas2r/esas2r_int.c 	if (likely(atomic_read(&a->disable_cnt) == 0))
atomic_read       471 drivers/scsi/esas2r/esas2r_int.c 	if (atomic_read(&a->disable_cnt) == 0)
atomic_read       661 drivers/scsi/esas2r/esas2r_int.c 	if (atomic_read(&a->disable_cnt) == 0)
atomic_read       754 drivers/scsi/esas2r/esas2r_io.c 	if (atomic_read(&a->disable_cnt) == 0)
atomic_read       835 drivers/scsi/esas2r/esas2r_io.c 	if (atomic_read(&a->disable_cnt) == 0)
atomic_read      1031 drivers/scsi/esas2r/esas2r_main.c 		if (atomic_read(&a->disable_cnt) == 0)
atomic_read       218 drivers/scsi/fnic/fnic_scsi.c 	while (atomic_read(&fnic->in_flight))
atomic_read       563 drivers/scsi/hosts.c 	return atomic_read(&shost->host_busy);
atomic_read       529 drivers/scsi/hpsa.c 			atomic_read(&h->commands_outstanding));
atomic_read      2453 drivers/scsi/hpsa.c 			atomic_read(&dev->commands_outstanding) <= 0)
atomic_read      3169 drivers/scsi/hpsa.c 			atomic_read(&dev->commands_outstanding) <= 0 ||
atomic_read      6039 drivers/scsi/hpsa.c 		if (atomic_read(&dev->commands_outstanding) > 0)
atomic_read      1082 drivers/scsi/hptiop.c 			atomic_read(&hba->resetting) == 0, 60 * HZ);
atomic_read      1084 drivers/scsi/hptiop.c 	if (atomic_read(&hba->resetting)) {
atomic_read      1243 drivers/scsi/ibmvscsi/ibmvfc.c 		BUG_ON(atomic_read(&pool->events[i].free) != 1);
atomic_read      2781 drivers/scsi/ibmvscsi/ibmvfc.c 	if (unlikely(atomic_read(&evt->free))) {
atomic_read       490 drivers/scsi/ibmvscsi/ibmvscsi.c 		if (atomic_read(&pool->events[i].free) != 1)
atomic_read      1730 drivers/scsi/ibmvscsi/ibmvscsi.c 		     atomic_read(&hostdata->request_limit) < 2;) {
atomic_read      1735 drivers/scsi/ibmvscsi/ibmvscsi.c 	if (atomic_read(&hostdata->request_limit) <= 0)
atomic_read      1814 drivers/scsi/ibmvscsi/ibmvscsi.c 	if (atomic_read(&evt_struct->free)) {
atomic_read      2288 drivers/scsi/ibmvscsi/ibmvscsi.c 		     atomic_read(&hostdata->request_limit) < 2;) {
atomic_read      2294 drivers/scsi/ibmvscsi/ibmvscsi.c 		if (atomic_read(&hostdata->request_limit) > 0)
atomic_read       117 drivers/scsi/iscsi_tcp.c 	    !atomic_read(&sk->sk_rmem_alloc)) {
atomic_read      2355 drivers/scsi/libfc/fc_exch.c 		st->fc_no_free_exch += atomic_read(&mp->stats.no_free_exch);
atomic_read      2357 drivers/scsi/libfc/fc_exch.c 				atomic_read(&mp->stats.no_free_exch_xid);
atomic_read      2358 drivers/scsi/libfc/fc_exch.c 		st->fc_xid_not_found += atomic_read(&mp->stats.xid_not_found);
atomic_read      2359 drivers/scsi/libfc/fc_exch.c 		st->fc_xid_busy += atomic_read(&mp->stats.xid_busy);
atomic_read      2360 drivers/scsi/libfc/fc_exch.c 		st->fc_seq_not_found += atomic_read(&mp->stats.seq_not_found);
atomic_read      2361 drivers/scsi/libfc/fc_exch.c 		st->fc_non_bls_resp += atomic_read(&mp->stats.non_bls_resp);
atomic_read       608 drivers/scsi/libsas/sas_init.c 	if (atomic_read(&phy->event_nr) > phy->ha->event_thres) {
atomic_read       220 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->rcv_ls_req_in),
atomic_read       221 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->rcv_ls_req_drop),
atomic_read       222 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_abort));
atomic_read       226 drivers/scsi/lpfc/lpfc_attr.c 		if (atomic_read(&tgtp->rcv_ls_req_in) !=
atomic_read       227 drivers/scsi/lpfc/lpfc_attr.c 		    atomic_read(&tgtp->rcv_ls_req_out)) {
atomic_read       230 drivers/scsi/lpfc/lpfc_attr.c 				  atomic_read(&tgtp->rcv_ls_req_in),
atomic_read       231 drivers/scsi/lpfc/lpfc_attr.c 				  atomic_read(&tgtp->rcv_ls_req_out));
atomic_read       238 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_rsp),
atomic_read       239 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_drop),
atomic_read       240 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_rsp_cmpl));
atomic_read       246 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_rsp_aborted),
atomic_read       247 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_rsp_xb_set),
atomic_read       248 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_ls_rsp_error));
atomic_read       255 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read       256 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->rcv_fcp_cmd_defer),
atomic_read       257 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_release),
atomic_read       258 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->rcv_fcp_cmd_drop));
atomic_read       262 drivers/scsi/lpfc/lpfc_attr.c 		if (atomic_read(&tgtp->rcv_fcp_cmd_in) !=
atomic_read       263 drivers/scsi/lpfc/lpfc_attr.c 		    atomic_read(&tgtp->rcv_fcp_cmd_out)) {
atomic_read       266 drivers/scsi/lpfc/lpfc_attr.c 				  atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read       267 drivers/scsi/lpfc/lpfc_attr.c 				  atomic_read(&tgtp->rcv_fcp_cmd_out));
atomic_read       275 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_read),
atomic_read       276 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_read_rsp),
atomic_read       277 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_write),
atomic_read       278 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_rsp),
atomic_read       279 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_drop));
atomic_read       285 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_rsp_cmpl),
atomic_read       286 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_rsp_error),
atomic_read       287 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_rsp_drop));
atomic_read       293 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_rsp_aborted),
atomic_read       294 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_rsp_xb_set),
atomic_read       295 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_xri_abort_cqe));
atomic_read       301 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_abort),
atomic_read       302 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_fcp_abort_cmpl));
atomic_read       308 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_abort_sol),
atomic_read       309 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_abort_unsol),
atomic_read       310 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_abort_rsp),
atomic_read       311 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->xmt_abort_rsp_error));
atomic_read       317 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->defer_ctx),
atomic_read       318 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->defer_fod),
atomic_read       319 drivers/scsi/lpfc/lpfc_attr.c 			  atomic_read(&tgtp->defer_wqfull));
atomic_read       324 drivers/scsi/lpfc/lpfc_attr.c 		tot = atomic_read(&tgtp->rcv_fcp_cmd_drop);
atomic_read       325 drivers/scsi/lpfc/lpfc_attr.c 		tot += atomic_read(&tgtp->xmt_fcp_release);
atomic_read       326 drivers/scsi/lpfc/lpfc_attr.c 		tot = atomic_read(&tgtp->rcv_fcp_cmd_in) - tot;
atomic_read       458 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->fc4NvmeLsRequests),
atomic_read       459 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->fc4NvmeLsCmpls),
atomic_read       460 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_ls_abort));
atomic_read       466 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_ls_err),
atomic_read       467 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->cmpl_ls_xb),
atomic_read       468 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->cmpl_ls_err));
atomic_read       493 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_fcp_abort),
atomic_read       494 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_fcp_noxri),
atomic_read       495 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_fcp_bad_ndlp),
atomic_read       496 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_fcp_qdepth),
atomic_read       497 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_fcp_err),
atomic_read       498 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->xmt_fcp_wqerr));
atomic_read       504 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->cmpl_fcp_xb),
atomic_read       505 drivers/scsi/lpfc/lpfc_attr.c 		  atomic_read(&lport->cmpl_fcp_err));
atomic_read       163 drivers/scsi/lpfc/lpfc_debugfs.c 	index = (atomic_read(&vport->disc_trc_cnt) + 1) &
atomic_read       229 drivers/scsi/lpfc/lpfc_debugfs.c 	index = (atomic_read(&phba->slow_ring_trc_cnt) + 1) &
atomic_read       905 drivers/scsi/lpfc/lpfc_debugfs.c 			i = atomic_read(&ndlp->cmd_pending);
atomic_read      1055 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->rcv_ls_req_in),
atomic_read      1056 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->rcv_ls_req_drop),
atomic_read      1057 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_abort));
atomic_read      1058 drivers/scsi/lpfc/lpfc_debugfs.c 		if (atomic_read(&tgtp->rcv_ls_req_in) !=
atomic_read      1059 drivers/scsi/lpfc/lpfc_debugfs.c 		    atomic_read(&tgtp->rcv_ls_req_out)) {
atomic_read      1062 drivers/scsi/lpfc/lpfc_debugfs.c 					atomic_read(&tgtp->rcv_ls_req_in),
atomic_read      1063 drivers/scsi/lpfc/lpfc_debugfs.c 					atomic_read(&tgtp->rcv_ls_req_out));
atomic_read      1068 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_rsp),
atomic_read      1069 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_drop),
atomic_read      1070 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_rsp_cmpl));
atomic_read      1074 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_rsp_aborted),
atomic_read      1075 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_rsp_xb_set),
atomic_read      1076 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_ls_rsp_error));
atomic_read      1081 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read      1082 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->rcv_fcp_cmd_defer),
atomic_read      1083 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_release),
atomic_read      1084 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->rcv_fcp_cmd_drop));
atomic_read      1086 drivers/scsi/lpfc/lpfc_debugfs.c 		if (atomic_read(&tgtp->rcv_fcp_cmd_in) !=
atomic_read      1087 drivers/scsi/lpfc/lpfc_debugfs.c 		    atomic_read(&tgtp->rcv_fcp_cmd_out)) {
atomic_read      1090 drivers/scsi/lpfc/lpfc_debugfs.c 					atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read      1091 drivers/scsi/lpfc/lpfc_debugfs.c 					atomic_read(&tgtp->rcv_fcp_cmd_out));
atomic_read      1097 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_read),
atomic_read      1098 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_read_rsp),
atomic_read      1099 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_write),
atomic_read      1100 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_rsp));
atomic_read      1104 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_rsp_cmpl),
atomic_read      1105 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_rsp_error),
atomic_read      1106 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_rsp_drop));
atomic_read      1110 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_rsp_aborted),
atomic_read      1111 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_rsp_xb_set),
atomic_read      1112 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_xri_abort_cqe));
atomic_read      1116 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_abort),
atomic_read      1117 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_fcp_abort_cmpl));
atomic_read      1121 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_abort_sol),
atomic_read      1122 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_abort_unsol),
atomic_read      1123 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_abort_rsp),
atomic_read      1124 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&tgtp->xmt_abort_rsp_error));
atomic_read      1155 drivers/scsi/lpfc/lpfc_debugfs.c 		tot = atomic_read(&tgtp->rcv_fcp_cmd_drop);
atomic_read      1156 drivers/scsi/lpfc/lpfc_debugfs.c 		tot += atomic_read(&tgtp->xmt_fcp_release);
atomic_read      1157 drivers/scsi/lpfc/lpfc_debugfs.c 		tot = atomic_read(&tgtp->rcv_fcp_cmd_in) - tot;
atomic_read      1182 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->fc4NvmeLsRequests),
atomic_read      1183 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->fc4NvmeLsCmpls));
atomic_read      1216 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_ls_abort),
atomic_read      1217 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_ls_err),
atomic_read      1218 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->cmpl_ls_xb),
atomic_read      1219 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->cmpl_ls_err));
atomic_read      1224 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_fcp_noxri),
atomic_read      1225 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_fcp_bad_ndlp),
atomic_read      1226 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_fcp_qdepth),
atomic_read      1227 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_fcp_wqerr),
atomic_read      1228 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_fcp_err),
atomic_read      1229 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->xmt_fcp_abort));
atomic_read      1233 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->cmpl_fcp_xb),
atomic_read      1234 drivers/scsi/lpfc/lpfc_debugfs.c 				atomic_read(&lport->cmpl_fcp_err));
atomic_read      1524 drivers/scsi/lpfc/lpfc_debugfs.c 	index = (atomic_read(&phba->nvmeio_trc_cnt) + 1) &
atomic_read      6112 drivers/scsi/lpfc/lpfc_debugfs.c 	if (atomic_read(&phba->debugfs_vport_count) == 0) {
atomic_read      6217 drivers/scsi/lpfc/lpfc_debugfs.c 		if (atomic_read(&lpfc_debugfs_hba_count) == 0) {
atomic_read      9610 drivers/scsi/lpfc/lpfc_els.c 	if (atomic_read(&phba->fabric_iocb_count) == 0) {
atomic_read      9730 drivers/scsi/lpfc/lpfc_els.c 	BUG_ON(atomic_read(&phba->fabric_iocb_count) == 0);
atomic_read      9775 drivers/scsi/lpfc/lpfc_els.c 	BUG_ON(atomic_read(&phba->fabric_iocb_count) > 1);
atomic_read      9778 drivers/scsi/lpfc/lpfc_els.c 	ready = atomic_read(&phba->fabric_iocb_count) == 0 &&
atomic_read       412 drivers/scsi/lpfc/lpfc_hbadisc.c 	if (atomic_read(&phba->fast_event_count) > LPFC_MAX_EVT_COUNT)
atomic_read      1611 drivers/scsi/lpfc/lpfc_nvme.c 		if ((atomic_read(&ndlp->cmd_pending) >= ndlp->cmd_qdepth) &&
atomic_read      1617 drivers/scsi/lpfc/lpfc_nvme.c 					 atomic_read(&ndlp->cmd_pending),
atomic_read       474 drivers/scsi/lpfc/lpfc_nvmet.c 					atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read       475 drivers/scsi/lpfc/lpfc_nvmet.c 					atomic_read(&tgtp->rcv_fcp_cmd_out),
atomic_read       476 drivers/scsi/lpfc/lpfc_nvmet.c 					atomic_read(&tgtp->xmt_fcp_release));
atomic_read      2133 drivers/scsi/lpfc/lpfc_nvmet.c 			atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read      2134 drivers/scsi/lpfc/lpfc_nvmet.c 			atomic_read(&tgtp->rcv_fcp_cmd_out),
atomic_read      2135 drivers/scsi/lpfc/lpfc_nvmet.c 			atomic_read(&tgtp->xmt_fcp_release));
atomic_read      2371 drivers/scsi/lpfc/lpfc_nvmet.c 				atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read      2372 drivers/scsi/lpfc/lpfc_nvmet.c 				atomic_read(&tgtp->rcv_fcp_cmd_out),
atomic_read      2373 drivers/scsi/lpfc/lpfc_nvmet.c 				atomic_read(&tgtp->xmt_fcp_release));
atomic_read       252 drivers/scsi/lpfc/lpfc_scsi.c 	num_rsrc_err = atomic_read(&phba->num_rsrc_err);
atomic_read       253 drivers/scsi/lpfc/lpfc_scsi.c 	num_cmd_success = atomic_read(&phba->num_cmd_success);
atomic_read      4017 drivers/scsi/lpfc/lpfc_scsi.c 				atomic_read(&pnode->cmd_pending) &&
atomic_read      4018 drivers/scsi/lpfc/lpfc_scsi.c 				(atomic_read(&pnode->cmd_pending) >
atomic_read      4023 drivers/scsi/lpfc/lpfc_scsi.c 					atomic_read(&pnode->cmd_pending);
atomic_read      4542 drivers/scsi/lpfc/lpfc_scsi.c 		if (atomic_read(&ndlp->cmd_pending) >= ndlp->cmd_qdepth) {
atomic_read      4551 drivers/scsi/lpfc/lpfc_scsi.c 					 atomic_read(&ndlp->cmd_pending),
atomic_read      13443 drivers/scsi/lpfc/lpfc_sli.c 					atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read      13444 drivers/scsi/lpfc/lpfc_sli.c 					atomic_read(&tgtp->rcv_fcp_cmd_out),
atomic_read      13445 drivers/scsi/lpfc/lpfc_sli.c 					atomic_read(&tgtp->xmt_fcp_release));
atomic_read      13953 drivers/scsi/lpfc/lpfc_sli.c 					atomic_read(&tgtp->rcv_fcp_cmd_in),
atomic_read      13954 drivers/scsi/lpfc/lpfc_sli.c 					atomic_read(&tgtp->rcv_fcp_cmd_out),
atomic_read      13955 drivers/scsi/lpfc/lpfc_sli.c 					atomic_read(&tgtp->xmt_fcp_release));
atomic_read       403 drivers/scsi/megaraid.c 	if (atomic_read(&adapter->quiescent) == 0)
atomic_read      1310 drivers/scsi/megaraid.c 		if(atomic_read(&adapter->quiescent) == 0) {
atomic_read      1390 drivers/scsi/megaraid.c 		if(atomic_read(&adapter->quiescent) == 0) {
atomic_read      2101 drivers/scsi/megaraid.c 		   atomic_read(&adapter->quiescent));
atomic_read      2126 drivers/scsi/megaraid.c 	seq_printf(m, "pend_cmds = %d\n", atomic_read(&adapter->pend_cmds));
atomic_read      3811 drivers/scsi/megaraid.c 	while (atomic_read(&adapter->pend_cmds) > 0 ||
atomic_read      4113 drivers/scsi/megaraid.c 	if (atomic_read(&adapter->quiescent) == 0)
atomic_read      4533 drivers/scsi/megaraid.c 	if (atomic_read(&adapter->pend_cmds) > 0)
atomic_read      3524 drivers/scsi/megaraid/megaraid_mbox.c 	if (atomic_read(&adapter->being_detached)) {
atomic_read       579 drivers/scsi/megaraid/megaraid_sas_base.c 	if ((atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) &&
atomic_read       717 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL)
atomic_read       856 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL)
atomic_read      1052 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL)
atomic_read      1099 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      1128 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      1195 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      1657 drivers/scsi/megaraid/megaraid_sas_base.c 	dev_err(&instance->pdev->dev, "[%d]: Total OS Pending cmds : %d\n",instance->host->host_no,atomic_read(&instance->fw_outstanding));
atomic_read      1777 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_ADPRESET_SM_INFAULT) {
atomic_read      1788 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      1801 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL)
atomic_read      2188 drivers/scsi/megaraid/megaraid_sas_base.c 	    && atomic_read(&instance->fw_outstanding) <
atomic_read      2216 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR)
atomic_read      2684 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      2690 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) {
atomic_read      2701 drivers/scsi/megaraid/megaraid_sas_base.c 			if (atomic_read(&instance->adprecovery) == MEGASAS_HBA_OPERATIONAL)
atomic_read      2705 drivers/scsi/megaraid/megaraid_sas_base.c 		if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) {
atomic_read      2745 drivers/scsi/megaraid/megaraid_sas_base.c 		outstanding = atomic_read(&instance->fw_outstanding);
atomic_read      2764 drivers/scsi/megaraid/megaraid_sas_base.c 	outstanding = atomic_read(&instance->fw_outstanding);
atomic_read      2773 drivers/scsi/megaraid/megaraid_sas_base.c 		if ((fw_state == MFI_STATE_FAULT) || atomic_read(&instance->fw_outstanding)) {
atomic_read      2776 drivers/scsi/megaraid/megaraid_sas_base.c 				__func__, __LINE__, fw_state, atomic_read(&instance->fw_outstanding));
atomic_read      2781 drivers/scsi/megaraid/megaraid_sas_base.c 			if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      2792 drivers/scsi/megaraid/megaraid_sas_base.c 			outstanding = atomic_read(&instance->fw_outstanding);
atomic_read      2813 drivers/scsi/megaraid/megaraid_sas_base.c 		atomic_read(&instance->fw_outstanding));
atomic_read      2838 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      3006 drivers/scsi/megaraid/megaraid_sas_base.c 		atomic_read(&instance->fw_outstanding));
atomic_read      3298 drivers/scsi/megaraid/megaraid_sas_base.c 	return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&instance->ldio_outstanding));
atomic_read      3308 drivers/scsi/megaraid/megaraid_sas_base.c 	return snprintf(buf, PAGE_SIZE, "%d\n", atomic_read(&instance->fw_outstanding));
atomic_read      3820 drivers/scsi/megaraid/megaraid_sas_base.c     if (atomic_read(&instance->adprecovery) != MEGASAS_ADPRESET_SM_INFAULT) {
atomic_read      3822 drivers/scsi/megaraid/megaraid_sas_base.c 				atomic_read(&instance->adprecovery));
atomic_read      3826 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_ADPRESET_SM_INFAULT) {
atomic_read      3940 drivers/scsi/megaraid/megaraid_sas_base.c 					fw_state, atomic_read(&instance->adprecovery));
atomic_read      3964 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->fw_reset_no_pci_access))
atomic_read      7442 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR)
atomic_read      7485 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR)
atomic_read      7733 drivers/scsi/megaraid/megaraid_sas_base.c 		adp_state = atomic_read(&instance->adprecovery);
atomic_read      7995 drivers/scsi/megaraid/megaraid_sas_base.c 			if ((atomic_read(&local_instance->adprecovery) ==
atomic_read      8283 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      8333 drivers/scsi/megaraid/megaraid_sas_base.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      1387 drivers/scsi/megaraid/megaraid_sas_fp.c 		pend0 = atomic_read(&lbInfo->scsi_pending_cmds[pd0]);
atomic_read      1388 drivers/scsi/megaraid/megaraid_sas_fp.c 		pend1 = atomic_read(&lbInfo->scsi_pending_cmds[pd1]);
atomic_read      2830 drivers/scsi/megaraid/megaraid_sas_fusion.c 		atomic_read(&scp->device->device_busy) >
atomic_read      3163 drivers/scsi/megaraid/megaraid_sas_fusion.c 		atomic_read(&scmd->device->device_busy) > MR_DEVICE_HIGH_IOPS_DEPTH)
atomic_read      3534 drivers/scsi/megaraid/megaraid_sas_fusion.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR)
atomic_read      3771 drivers/scsi/megaraid/megaraid_sas_fusion.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR)
atomic_read      4173 drivers/scsi/megaraid/megaraid_sas_fusion.c 		outstanding = atomic_read(&instance->fw_outstanding);
atomic_read      4191 drivers/scsi/megaraid/megaraid_sas_fusion.c 	if (atomic_read(&instance->fw_outstanding)) {
atomic_read      4624 drivers/scsi/megaraid/megaraid_sas_fusion.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) {
atomic_read      4705 drivers/scsi/megaraid/megaraid_sas_fusion.c 	if (atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) {
atomic_read      4781 drivers/scsi/megaraid/megaraid_sas_fusion.c 			(atomic_read(&peer_instance->adprecovery) ==
atomic_read      4809 drivers/scsi/megaraid/megaraid_sas_fusion.c 	if (atomic_read(&instance->adprecovery) == MEGASAS_HW_CRITICAL_ERROR) {
atomic_read      4836 drivers/scsi/megaraid/megaraid_sas_fusion.c 		} while ((atomic_read(&instance->adprecovery) != MEGASAS_HBA_OPERATIONAL) &&
atomic_read      4839 drivers/scsi/megaraid/megaraid_sas_fusion.c 		if (atomic_read(&instance->adprecovery) == MEGASAS_HBA_OPERATIONAL) {
atomic_read      1838 drivers/scsi/mpt3sas/mpt3sas_base.c 	   atomic_read(&ioc->chain_lookup[smid - 1].chain_offset);
atomic_read      3488 drivers/scsi/mpt3sas/mpt3sas_base.c 	if (atomic_read(&scmd->device->device_busy) >
atomic_read      2996 drivers/scsi/mpt3sas/mpt3sas_scsih.c 	if (r == SUCCESS && atomic_read(&scmd->device->device_busy))
atomic_read      3074 drivers/scsi/mpt3sas/mpt3sas_scsih.c 	if (r == SUCCESS && atomic_read(&starget->target_busy))
atomic_read       416 drivers/scsi/mvumi.c 	if (atomic_read(&mhba->fw_outstanding) >= mhba->max_io) {
atomic_read       420 drivers/scsi/mvumi.c 		return mhba->max_io - atomic_read(&mhba->fw_outstanding);
atomic_read       427 drivers/scsi/mvumi.c 	if (atomic_read(&mhba->fw_outstanding) >= (mhba->max_io - 1))
atomic_read       717 drivers/scsi/mvumi.c 	if (atomic_read(&cmd->sync_cmd)) {
atomic_read       720 drivers/scsi/mvumi.c 						atomic_read(&cmd->sync_cmd));
atomic_read       734 drivers/scsi/mvumi.c 	if (atomic_read(&cmd->sync_cmd)) {
atomic_read      1339 drivers/scsi/mvumi.c 	if (atomic_read(&cmd->sync_cmd)) {
atomic_read      1614 drivers/scsi/mvumi.c 		if (!atomic_read(&mhba->pnp_count))
atomic_read      1761 drivers/scsi/pmcraid.c 	    atomic_read(&pinstance->ccn.ignore) == 1) {
atomic_read      1802 drivers/scsi/pmcraid.c 	    atomic_read(&pinstance->ccn.ignore) == 1) {
atomic_read      3076 drivers/scsi/pmcraid.c 		if (atomic_read(&pinstance->outstanding_cmds) <=
atomic_read      4302 drivers/scsi/pmcraid.c 	if (!atomic_read(&pinstance->expose_resources))
atomic_read      5660 drivers/scsi/pmcraid.c 	if (atomic_read(&pmcraid_adapter_count) >= PMCRAID_MAX_ADAPTERS) {
atomic_read      5663 drivers/scsi/pmcraid.c 			 atomic_read(&pmcraid_adapter_count));
atomic_read      5679 drivers/scsi/pmcraid.c 		 atomic_read(&pmcraid_adapter_count));
atomic_read       330 drivers/scsi/qedf/qedf_debugfs.c 	seq_printf(s, "Link State: %s\n", atomic_read(&qedf->link_state) ?
atomic_read       347 drivers/scsi/qedf/qedf_debugfs.c 	    atomic_read(&qedf->cmd_mgr->free_list_cnt));
atomic_read       362 drivers/scsi/qedf/qedf_debugfs.c 			   atomic_read(&fcport->free_sqes),
atomic_read       363 drivers/scsi/qedf/qedf_debugfs.c 			   atomic_read(&fcport->num_active_ios));
atomic_read        61 drivers/scsi/qedf/qedf_fip.c 	if (atomic_read(&qedf->link_state) != QEDF_LINK_UP) {
atomic_read       103 drivers/scsi/qedf/qedf_fip.c 	if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) {
atomic_read       296 drivers/scsi/qedf/qedf_io.c 	    atomic_read(&cmgr->free_list_cnt));
atomic_read       316 drivers/scsi/qedf/qedf_io.c 	free_sqes = atomic_read(&fcport->free_sqes);
atomic_read       326 drivers/scsi/qedf/qedf_io.c 	if ((atomic_read(&fcport->num_active_ios) >=
atomic_read       330 drivers/scsi/qedf/qedf_io.c 		    atomic_read(&fcport->num_active_ios));
atomic_read       335 drivers/scsi/qedf/qedf_io.c 	if (atomic_read(&cmd_mgr->free_list_cnt) <= GBL_RSVD_TASKS) {
atomic_read       338 drivers/scsi/qedf/qedf_io.c 		    atomic_read(&cmd_mgr->free_list_cnt));
atomic_read       461 drivers/scsi/qedf/qedf_io.c 	if (atomic_read(&fcport->num_active_ios) < 0) {
atomic_read      1002 drivers/scsi/qedf/qedf_io.c 	    atomic_read(&qedf->link_state) != QEDF_LINK_UP) {
atomic_read      1597 drivers/scsi/qedf/qedf_io.c 		while (atomic_read(&fcport->ios_to_queue)) {
atomic_read      1600 drivers/scsi/qedf/qedf_io.c 				  atomic_read(&fcport->ios_to_queue));
atomic_read      1604 drivers/scsi/qedf/qedf_io.c 					 atomic_read(&fcport->ios_to_queue));
atomic_read      1615 drivers/scsi/qedf/qedf_io.c 		  atomic_read(&fcport->num_active_ios), fcport,
atomic_read      1666 drivers/scsi/qedf/qedf_io.c 			if (atomic_read(&io_req->state) ==
atomic_read      1781 drivers/scsi/qedf/qedf_io.c 		  flush_cnt, atomic_read(&fcport->num_active_ios));
atomic_read      1785 drivers/scsi/qedf/qedf_io.c 		while (atomic_read(&fcport->num_active_ios)) {
atomic_read      1789 drivers/scsi/qedf/qedf_io.c 				  atomic_read(&fcport->num_active_ios),
atomic_read      1795 drivers/scsi/qedf/qedf_io.c 					 atomic_read(&fcport->num_active_ios));
atomic_read      1870 drivers/scsi/qedf/qedf_io.c 	if (atomic_read(&qedf->link_down_tmo_valid) > 0) {
atomic_read      1877 drivers/scsi/qedf/qedf_io.c 	if (!atomic_read(&fcport->free_sqes)) {
atomic_read      2172 drivers/scsi/qedf/qedf_io.c 	if (!atomic_read(&fcport->free_sqes)) {
atomic_read       129 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) {
atomic_read       143 drivers/scsi/qedf/qedf_main.c 			if (atomic_read(&qedf->link_state) == QEDF_LINK_UP)
atomic_read       165 drivers/scsi/qedf/qedf_main.c 		  atomic_read(&qedf->link_state));
atomic_read       167 drivers/scsi/qedf/qedf_main.c 	if (atomic_read(&qedf->link_state) == QEDF_LINK_UP) {
atomic_read       172 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->link_state) != QEDF_LINK_UP) {
atomic_read       195 drivers/scsi/qedf/qedf_main.c 	} else if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) {
atomic_read       521 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->link_state) == QEDF_LINK_UP) {
atomic_read       535 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->dcbx) == QEDF_DCBX_DONE ||
atomic_read       539 drivers/scsi/qedf/qedf_main.c 			if (atomic_read(&qedf->link_down_tmo_valid) > 0)
atomic_read       581 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->dcbx) == QEDF_DCBX_DONE) {
atomic_read       607 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->link_state) == QEDF_LINK_UP &&
atomic_read       609 drivers/scsi/qedf/qedf_main.c 			if (atomic_read(&qedf->link_down_tmo_valid) > 0)
atomic_read       813 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->num_offloads))
atomic_read       816 drivers/scsi/qedf/qedf_main.c 				  atomic_read(&qedf->num_offloads));
atomic_read       862 drivers/scsi/qedf/qedf_main.c 		WARN_ON(atomic_read(&qedf->num_offloads));
atomic_read       894 drivers/scsi/qedf/qedf_main.c 	if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN ||
atomic_read      1060 drivers/scsi/qedf/qedf_main.c 	if (atomic_read(&qedf->link_state) != QEDF_LINK_UP) {
atomic_read      1412 drivers/scsi/qedf/qedf_main.c 		if (atomic_read(&qedf->num_offloads) >= QEDF_MAX_SESSIONS) {
atomic_read      1735 drivers/scsi/qedf/qedf_main.c 	if (atomic_read(&base_qedf->link_state) != QEDF_LINK_UP) {
atomic_read      2529 drivers/scsi/qedf/qedf_main.c 	if (atomic_read(&qedf->link_state) == QEDF_LINK_DOWN) {
atomic_read      3668 drivers/scsi/qedf/qedf_main.c 	if (!atomic_read(&pdev->enable_cnt))
atomic_read       869 drivers/scsi/qedi/qedi_iscsi.c 	if (atomic_read(&qedi->link_state) != QEDI_LINK_UP) {
atomic_read        25 drivers/scsi/qedi/qedi_sysfs.c 	if (atomic_read(&qedi->link_state) == QEDI_LINK_UP)
atomic_read      1140 drivers/scsi/qla2xxx/qla_attr.c 	if (atomic_read(&vha->loop_state) == LOOP_DOWN ||
atomic_read      1141 drivers/scsi/qla2xxx/qla_attr.c 	    atomic_read(&vha->loop_state) == LOOP_DEAD ||
atomic_read      1144 drivers/scsi/qla2xxx/qla_attr.c 	else if (atomic_read(&vha->loop_state) != LOOP_READY ||
atomic_read      2636 drivers/scsi/qla2xxx/qla_attr.c 	} else if (atomic_read(&base_vha->loop_state) == LOOP_READY &&
atomic_read      2753 drivers/scsi/qla2xxx/qla_attr.c 	switch (atomic_read(&base_vha->loop_state)) {
atomic_read      2814 drivers/scsi/qla2xxx/qla_attr.c 	if (atomic_read(&base_vha->loop_state) == LOOP_DOWN ||
atomic_read      2815 drivers/scsi/qla2xxx/qla_attr.c 	    atomic_read(&base_vha->loop_state) == LOOP_DEAD) {
atomic_read       797 drivers/scsi/qla2xxx/qla_bsg.c 	if (atomic_read(&vha->loop_state) == LOOP_READY &&
atomic_read      1333 drivers/scsi/qla2xxx/qla_bsg.c 	if (atomic_read(&fcport->state) != FCS_ONLINE) {
atomic_read      4616 drivers/scsi/qla2xxx/qla_def.h 	 atomic_read(&ha->loop_state) == LOOP_DOWN)
atomic_read       525 drivers/scsi/qla2xxx/qla_dfs.c 	if (atomic_read(&qla2x00_dfs_root_count) == 0 &&
atomic_read      3677 drivers/scsi/qla2xxx/qla_gs.c 			    atomic_read(&fcport->state) == FCS_ONLINE) {
atomic_read      4326 drivers/scsi/qla2xxx/qla_gs.c 	ls = atomic_read(&vha->loop_state);
atomic_read      4455 drivers/scsi/qla2xxx/qla_gs.c 	ls = atomic_read(&vha->loop_state);
atomic_read      4237 drivers/scsi/qla2xxx/qla_init.c 			if (atomic_read(&vha->loop_down_timer) &&
atomic_read      4309 drivers/scsi/qla2xxx/qla_init.c 		if (LOOP_TRANSITION(vha) || atomic_read(&ha->loop_down_timer) ||
atomic_read      4807 drivers/scsi/qla2xxx/qla_init.c 	old_state = atomic_read(&fcport->state);
atomic_read      4978 drivers/scsi/qla2xxx/qla_init.c 		if (atomic_read(&vha->loop_down_timer) ||
atomic_read      5240 drivers/scsi/qla2xxx/qla_init.c 			    atomic_read(&fcport->state) == FCS_ONLINE) {
atomic_read      5283 drivers/scsi/qla2xxx/qla_init.c 	if (atomic_read(&fcport->state) != FCS_ONLINE)
atomic_read      5331 drivers/scsi/qla2xxx/qla_init.c 	if (atomic_read(&fcport->state) == FCS_ONLINE)
atomic_read      5720 drivers/scsi/qla2xxx/qla_init.c 		    (atomic_read(&vha->loop_down_timer) ||
atomic_read      5830 drivers/scsi/qla2xxx/qla_init.c 			    (atomic_read(&fcport->state) == FCS_ONLINE ||
atomic_read      5918 drivers/scsi/qla2xxx/qla_init.c 			    atomic_read(&fcport->state) == FCS_ONLINE) {
atomic_read      6189 drivers/scsi/qla2xxx/qla_init.c 			} while (!atomic_read(&vha->loop_down_timer) &&
atomic_read      6251 drivers/scsi/qla2xxx/qla_init.c 			    atomic_read(&fcport->state) != FCS_UNCONFIGURED) {
atomic_read      6598 drivers/scsi/qla2xxx/qla_init.c 	if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read      6604 drivers/scsi/qla2xxx/qla_init.c 		if (!atomic_read(&vha->loop_down_timer))
atomic_read      6659 drivers/scsi/qla2xxx/qla_init.c 	if (atomic_read(&ha->num_pend_mbx_stage3)) {
atomic_read      6665 drivers/scsi/qla2xxx/qla_init.c 	while (atomic_read(&ha->num_pend_mbx_stage3) ||
atomic_read      6666 drivers/scsi/qla2xxx/qla_init.c 	    atomic_read(&ha->num_pend_mbx_stage2) ||
atomic_read      6667 drivers/scsi/qla2xxx/qla_init.c 	    atomic_read(&ha->num_pend_mbx_stage1)) {
atomic_read      6676 drivers/scsi/qla2xxx/qla_init.c 	if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read      6692 drivers/scsi/qla2xxx/qla_init.c 		if (!atomic_read(&vha->loop_down_timer))
atomic_read      6802 drivers/scsi/qla2xxx/qla_init.c 			if (!atomic_read(&vha->loop_down_timer)) {
atomic_read      8583 drivers/scsi/qla2xxx/qla_init.c 		if (!atomic_read(&vha->loop_down_timer)) {
atomic_read       788 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read       838 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read       881 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read       921 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read       923 drivers/scsi/qla2xxx/qla_isr.c 			if (!atomic_read(&vha->loop_down_timer))
atomic_read       951 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read       953 drivers/scsi/qla2xxx/qla_isr.c 			if (!atomic_read(&vha->loop_down_timer))
atomic_read      1019 drivers/scsi/qla2xxx/qla_isr.c 			if (atomic_read(&fcport->state) != FCS_ONLINE)
atomic_read      1032 drivers/scsi/qla2xxx/qla_isr.c 			if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read      1058 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&vha->loop_state) != LOOP_DOWN &&
atomic_read      1060 drivers/scsi/qla2xxx/qla_isr.c 		    atomic_read(&vha->loop_state) != LOOP_DEAD) {
atomic_read      1208 drivers/scsi/qla2xxx/qla_isr.c 				if (atomic_read(&vha->loop_state) == LOOP_DOWN)
atomic_read      2731 drivers/scsi/qla2xxx/qla_isr.c 		if (atomic_read(&fcport->state) == FCS_ONLINE) {
atomic_read        81 drivers/scsi/qla2xxx/qla_mid.c 		    !atomic_read(&vha->vref_count), HZ) > 0)
atomic_read        86 drivers/scsi/qla2xxx/qla_mid.c 	if (atomic_read(&vha->vref_count)) {
atomic_read       197 drivers/scsi/qla2xxx/qla_mid.c 	if (atomic_read(&base_vha->loop_state) == LOOP_DOWN  ||
atomic_read       198 drivers/scsi/qla2xxx/qla_mid.c 		atomic_read(&base_vha->loop_state) == LOOP_DEAD ||
atomic_read       328 drivers/scsi/qla2xxx/qla_mid.c 	if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read       332 drivers/scsi/qla2xxx/qla_mid.c 		if (!atomic_read(&vha->loop_down_timer))
atomic_read       375 drivers/scsi/qla2xxx/qla_mid.c 	    atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read      1133 drivers/scsi/qla2xxx/qla_mr.c 	if ((atomic_read(&vha->loop_down_timer) ||
atomic_read      1178 drivers/scsi/qla2xxx/qla_mr.c 			    atomic_read(&fcport->state) == FCS_ONLINE)
atomic_read      1189 drivers/scsi/qla2xxx/qla_mr.c 			    atomic_read(&fcport->state),
atomic_read      1202 drivers/scsi/qla2xxx/qla_mr.c 			if (atomic_read(&fcport->state) != FCS_ONLINE) {
atomic_read      1275 drivers/scsi/qla2xxx/qla_mr.c 		if (atomic_read(&fcport->state) == FCS_DEVICE_LOST) {
atomic_read      1372 drivers/scsi/qla2xxx/qla_mr.c 	if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
atomic_read      1377 drivers/scsi/qla2xxx/qla_mr.c 		if (!atomic_read(&vha->loop_down_timer))
atomic_read      1385 drivers/scsi/qla2xxx/qla_mr.c 		if (atomic_read(&fcport->state) == FCS_ONLINE)
atomic_read      2513 drivers/scsi/qla2xxx/qla_mr.c 		    atomic_read(&fcport->state));
atomic_read      2515 drivers/scsi/qla2xxx/qla_mr.c 		if (atomic_read(&fcport->state) == FCS_ONLINE)
atomic_read       867 drivers/scsi/qla2xxx/qla_os.c 	if (atomic_read(&fcport->state) != FCS_ONLINE) {
atomic_read       868 drivers/scsi/qla2xxx/qla_os.c 		if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD ||
atomic_read       869 drivers/scsi/qla2xxx/qla_os.c 			atomic_read(&base_vha->loop_state) == LOOP_DEAD) {
atomic_read       872 drivers/scsi/qla2xxx/qla_os.c 			    atomic_read(&fcport->state),
atomic_read       873 drivers/scsi/qla2xxx/qla_os.c 			    atomic_read(&base_vha->loop_state));
atomic_read       949 drivers/scsi/qla2xxx/qla_os.c 	if (atomic_read(&fcport->state) != FCS_ONLINE) {
atomic_read       950 drivers/scsi/qla2xxx/qla_os.c 		if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD ||
atomic_read       951 drivers/scsi/qla2xxx/qla_os.c 			atomic_read(&base_vha->loop_state) == LOOP_DEAD) {
atomic_read       954 drivers/scsi/qla2xxx/qla_os.c 			    atomic_read(&fcport->state),
atomic_read       955 drivers/scsi/qla2xxx/qla_os.c 			    atomic_read(&base_vha->loop_state));
atomic_read      2723 drivers/scsi/qla2xxx/qla_os.c 	return atomic_read(&vha->loop_state) == LOOP_READY;
atomic_read      3524 drivers/scsi/qla2xxx/qla_os.c 	if (!atomic_read(&pdev->enable_cnt))
atomic_read      3692 drivers/scsi/qla2xxx/qla_os.c 	if (!atomic_read(&pdev->enable_cnt)) {
atomic_read      3886 drivers/scsi/qla2xxx/qla_os.c 	if (atomic_read(&fcport->state) == FCS_ONLINE &&
atomic_read      3895 drivers/scsi/qla2xxx/qla_os.c 	if (atomic_read(&fcport->state) != FCS_DEVICE_DEAD)
atomic_read      3936 drivers/scsi/qla2xxx/qla_os.c 		if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD)
atomic_read      3938 drivers/scsi/qla2xxx/qla_os.c 		if (atomic_read(&fcport->state) == FCS_ONLINE) {
atomic_read      5359 drivers/scsi/qla2xxx/qla_os.c 		if (atomic_read(&fcport->state) != FCS_ONLINE &&
atomic_read      6060 drivers/scsi/qla2xxx/qla_os.c 	if (!atomic_read(&pdev->enable_cnt)) {
atomic_read      6368 drivers/scsi/qla2xxx/qla_os.c 		    atomic_read(&base_vha->loop_state) != LOOP_DOWN) {
atomic_read      6404 drivers/scsi/qla2xxx/qla_os.c 		    atomic_read(&base_vha->loop_state) == LOOP_READY) {
atomic_read      6506 drivers/scsi/qla2xxx/qla_os.c 	    !atomic_read(&vha->loop_down_timer) &&
atomic_read      6516 drivers/scsi/qla2xxx/qla_os.c 		} while (!atomic_read(&vha->loop_down_timer) &&
atomic_read      6572 drivers/scsi/qla2xxx/qla_os.c 	if (atomic_read(&vha->loop_down_timer) > 0 &&
atomic_read      6577 drivers/scsi/qla2xxx/qla_os.c 		if (atomic_read(&vha->loop_down_timer) ==
atomic_read      6641 drivers/scsi/qla2xxx/qla_os.c 		    atomic_read(&vha->loop_down_timer));
atomic_read      6670 drivers/scsi/qla2xxx/qla_os.c 	    (atomic_read(&ha->nvme_active_aen_cnt) != ha->nvme_last_rptd_aen) &&
atomic_read      6676 drivers/scsi/qla2xxx/qla_os.c 		ha->nvme_last_rptd_aen = atomic_read(&ha->nvme_active_aen_cnt);
atomic_read      6682 drivers/scsi/qla2xxx/qla_os.c 	    (atomic_read(&ha->zio_threshold) != ha->last_zio_threshold) &&
atomic_read      6688 drivers/scsi/qla2xxx/qla_os.c 		ha->last_zio_threshold = atomic_read(&ha->zio_threshold);
atomic_read      6859 drivers/scsi/qla2xxx/qla_os.c 	if (atomic_read(&ha->num_pend_mbx_stage3)) {
atomic_read      6866 drivers/scsi/qla2xxx/qla_os.c 	while (atomic_read(&ha->num_pend_mbx_stage3) ||
atomic_read      6867 drivers/scsi/qla2xxx/qla_os.c 	    atomic_read(&ha->num_pend_mbx_stage2) ||
atomic_read      6868 drivers/scsi/qla2xxx/qla_os.c 	    atomic_read(&ha->num_pend_mbx_stage1)) {
atomic_read      6920 drivers/scsi/qla2xxx/qla_os.c 	if (!atomic_read(&pdev->enable_cnt)) {
atomic_read      6114 drivers/scsi/qla2xxx/qla_target.c 	    atomic_read(&vha->vha_tgt.qla_tgt->tgt_global_resets_count);
atomic_read      6143 drivers/scsi/qla2xxx/qla_target.c 	    atomic_read(&vha->vha_tgt.qla_tgt->tgt_global_resets_count)) {
atomic_read      6148 drivers/scsi/qla2xxx/qla_target.c 		    atomic_read(&vha->vha_tgt.
atomic_read       892 drivers/scsi/qla2xxx/tcm_qla2xxx.c 			atomic_read(&tpg->lport_tpg_enabled));
atomic_read       918 drivers/scsi/qla2xxx/tcm_qla2xxx.c 		if (atomic_read(&tpg->lport_tpg_enabled))
atomic_read       924 drivers/scsi/qla2xxx/tcm_qla2xxx.c 		if (!atomic_read(&tpg->lport_tpg_enabled))
atomic_read      1080 drivers/scsi/qla2xxx/tcm_qla2xxx.c 		if (atomic_read(&tpg->lport_tpg_enabled))
atomic_read      1086 drivers/scsi/qla2xxx/tcm_qla2xxx.c 		if (!atomic_read(&tpg->lport_tpg_enabled))
atomic_read      1691 drivers/scsi/qla2xxx/tcm_qla2xxx.c 	    !atomic_read(&base_lport->tpg_1->lport_tpg_enabled)) {
atomic_read      4477 drivers/scsi/qla4xxx/ql4_os.c 		if (atomic_read(&ddb_entry->retry_relogin_timer) !=
atomic_read      4479 drivers/scsi/qla4xxx/ql4_os.c 			if (atomic_read(&ddb_entry->retry_relogin_timer) ==
atomic_read      4494 drivers/scsi/qla4xxx/ql4_os.c 	if (atomic_read(&ddb_entry->relogin_timer) &&
atomic_read      4507 drivers/scsi/qla4xxx/ql4_os.c 				atomic_read(&ddb_entry->relogin_retry_count),
atomic_read      8893 drivers/scsi/qla4xxx/ql4_os.c 		if (atomic_read(&other_pdev->enable_cnt)) {
atomic_read      9664 drivers/scsi/qla4xxx/ql4_os.c 			if (atomic_read(&other_pdev->enable_cnt)) {
atomic_read       195 drivers/scsi/scsi.c 	if (atomic_read(&shost->host_blocked))
atomic_read       197 drivers/scsi/scsi.c 	if (atomic_read(&starget->target_blocked))
atomic_read       199 drivers/scsi/scsi.c 	if (atomic_read(&sdev->device_blocked))
atomic_read      1617 drivers/scsi/scsi_debug.c 	changing = atomic_read(&devip->stopped) == !stop;
atomic_read      3747 drivers/scsi/scsi_debug.c 	if (unlikely(atomic_read(&retired_max_queue) > 0))
atomic_read      3760 drivers/scsi/scsi_debug.c 		retval = atomic_read(&retired_max_queue);
atomic_read      3935 drivers/scsi/scsi_debug.c 		r_qmax = atomic_read(&retired_max_queue);
atomic_read      4216 drivers/scsi/scsi_debug.c 	count = atomic_read(&sdebug_cmnd_count);
atomic_read      4232 drivers/scsi/scsi_debug.c 	if ((atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth)) > 0) {
atomic_read      4280 drivers/scsi/scsi_debug.c 	if (unlikely(atomic_read(&sqp->blocked))) {
atomic_read      4284 drivers/scsi/scsi_debug.c 	num_in_q = atomic_read(&devip->num_in_q);
atomic_read      4598 drivers/scsi/scsi_debug.c 		   atomic_read(&sdebug_cmnd_count),
atomic_read      4599 drivers/scsi/scsi_debug.c 		   atomic_read(&sdebug_completions),
atomic_read      4600 drivers/scsi/scsi_debug.c 		   "miss_cpus", atomic_read(&sdebug_miss_cpus),
atomic_read      4601 drivers/scsi/scsi_debug.c 		   atomic_read(&sdebug_a_tsf));
atomic_read      5552 drivers/scsi/scsi_debug.c 	num_in_q = atomic_read(&devip->num_in_q);
atomic_read      5571 drivers/scsi/scsi_debug.c 	if (0 == (atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth))) {
atomic_read      5586 drivers/scsi/scsi_debug.c 		(atomic_read(&sdebug_cmnd_count) % abs(sdebug_every_nth)) == 0;
atomic_read      5705 drivers/scsi/scsi_debug.c 	if (unlikely((F_M_ACCESS & flags) && atomic_read(&devip->stopped))) {
atomic_read       413 drivers/scsi/scsi_lib.c 	if (atomic_read(&sdev->device_busy) >= sdev->queue_depth)
atomic_read       415 drivers/scsi/scsi_lib.c 	if (atomic_read(&sdev->device_blocked) > 0)
atomic_read       423 drivers/scsi/scsi_lib.c 		if (atomic_read(&starget->target_busy) >= starget->can_queue)
atomic_read       425 drivers/scsi/scsi_lib.c 		if (atomic_read(&starget->target_blocked) > 0)
atomic_read       434 drivers/scsi/scsi_lib.c 	    atomic_read(&shost->host_busy) >= shost->can_queue)
atomic_read       436 drivers/scsi/scsi_lib.c 	if (atomic_read(&shost->host_blocked) > 0)
atomic_read      1287 drivers/scsi/scsi_lib.c 	if (atomic_read(&sdev->device_blocked)) {
atomic_read      1334 drivers/scsi/scsi_lib.c 	if (atomic_read(&starget->target_blocked) > 0) {
atomic_read      1378 drivers/scsi/scsi_lib.c 	if (atomic_read(&shost->host_blocked) > 0) {
atomic_read      1639 drivers/scsi/scsi_lib.c 	if (atomic_read(&sdev->device_busy) == 0 && !scsi_device_blocked(sdev))
atomic_read      1710 drivers/scsi/scsi_lib.c 		if (atomic_read(&sdev->device_busy) ||
atomic_read      2915 drivers/scsi/scsi_lib.c 	if (WARN_ON_ONCE(atomic_read(&sdev->disk_events_disable_depth) <= 0))
atomic_read       653 drivers/scsi/scsi_sysfs.c 	return snprintf(buf, 20, "%d\n", atomic_read(&sdev->device_busy));
atomic_read       662 drivers/scsi/scsi_sysfs.c 	return snprintf(buf, 20, "%d\n", atomic_read(&sdev->device_blocked));
atomic_read       910 drivers/scsi/scsi_sysfs.c 	unsigned long long count = atomic_read(&sdev->field);		\
atomic_read      1775 drivers/scsi/scsi_transport_iscsi.c 	return !atomic_read(&ihost->nr_scans);
atomic_read       254 drivers/scsi/sg.c 					(atomic_read(&sdp->detaching) ||
atomic_read       260 drivers/scsi/sg.c 			if (atomic_read(&sdp->detaching))
atomic_read       267 drivers/scsi/sg.c 					(atomic_read(&sdp->detaching) ||
atomic_read       273 drivers/scsi/sg.c 			if (atomic_read(&sdp->detaching))
atomic_read       464 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching)) {
atomic_read       473 drivers/scsi/sg.c 			(atomic_read(&sdp->detaching) ||
atomic_read       475 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching)) {
atomic_read       624 drivers/scsi/sg.c 	if (atomic_read(&sdp->detaching))
atomic_read       821 drivers/scsi/sg.c 	if (atomic_read(&sdp->detaching)) {
atomic_read       920 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching))
atomic_read       931 drivers/scsi/sg.c 			(srp_done(sfp, srp) || atomic_read(&sdp->detaching)));
atomic_read       932 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching))
atomic_read       975 drivers/scsi/sg.c 			if (atomic_read(&sdp->detaching))
atomic_read      1099 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching))
atomic_read      1103 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching))
atomic_read      1131 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching))
atomic_read      1196 drivers/scsi/sg.c 	if (atomic_read(&sdp->detaching))
atomic_read      1339 drivers/scsi/sg.c 	if (unlikely(atomic_read(&sdp->detaching)))
atomic_read      2169 drivers/scsi/sg.c 	if (atomic_read(&sdp->detaching)) {
atomic_read      2285 drivers/scsi/sg.c 	else if (atomic_read(&sdp->detaching)) {
atomic_read      2481 drivers/scsi/sg.c 	    (atomic_read(&sdp->detaching)))
atomic_read      2490 drivers/scsi/sg.c 			      (int) atomic_read(&scsidp->device_busy),
atomic_read      2507 drivers/scsi/sg.c 	if (sdp && scsidp && (!atomic_read(&sdp->detaching)))
atomic_read      2597 drivers/scsi/sg.c 		if (atomic_read(&sdp->detaching))
atomic_read       288 drivers/scsi/smartpqi/smartpqi_init.c 	while (atomic_read(&ctrl_info->num_busy_threads) >
atomic_read       289 drivers/scsi/smartpqi/smartpqi_init.c 		atomic_read(&ctrl_info->num_blocked_threads))
atomic_read      1515 drivers/scsi/smartpqi/smartpqi_init.c 			atomic_read(&device->scsi_cmds_outstanding));
atomic_read      2932 drivers/scsi/smartpqi/smartpqi_init.c 		WARN_ON(atomic_read(&io_request->refcount) == 0);
atomic_read      3218 drivers/scsi/smartpqi/smartpqi_init.c 	num_interrupts = atomic_read(&ctrl_info->num_interrupts);
atomic_read      3244 drivers/scsi/smartpqi/smartpqi_init.c 		atomic_read(&ctrl_info->num_interrupts);
atomic_read      5535 drivers/scsi/smartpqi/smartpqi_init.c 	while (atomic_read(&device->scsi_cmds_outstanding)) {
atomic_read      5567 drivers/scsi/smartpqi/smartpqi_init.c 			if (atomic_read(&device->scsi_cmds_outstanding)) {
atomic_read      5597 drivers/scsi/smartpqi/smartpqi_init.c 	while (atomic_read(&ctrl_info->sync_cmds_outstanding)) {
atomic_read      7793 drivers/scsi/smartpqi/smartpqi_init.c 		if (atomic_read(&io_request->refcount) == 0)
atomic_read      7934 drivers/scsi/smartpqi/smartpqi_init.c 		if (atomic_read(&io_request->refcount) == 0)
atomic_read       343 drivers/scsi/snic/snic_main.c 	return atomic_read(&snic->state);
atomic_read      2355 drivers/scsi/snic/snic_scsi.c 	while (atomic_read(&snic->ios_inflight))
atomic_read       611 drivers/scsi/sr.c 	if (!atomic_read(&cd->device->disk_events_disable_depth))
atomic_read       486 drivers/scsi/st.c 			atomic64_add(atomic_read(&STp->stats->last_write_size)
atomic_read       492 drivers/scsi/st.c 			atomic64_add(atomic_read(&STp->stats->last_write_size),
atomic_read       500 drivers/scsi/st.c 			atomic64_add(atomic_read(&STp->stats->last_read_size)
atomic_read       506 drivers/scsi/st.c 			atomic64_add(atomic_read(&STp->stats->last_read_size),
atomic_read       596 drivers/scsi/storvsc_drv.c 		   atomic_read(&dev->num_outstanding_req) == 0);
atomic_read       616 drivers/scsi/storvsc_drv.c 		(atomic_read(&stor_device->num_outstanding_req) == 0))
atomic_read       423 drivers/sh/maple/maple.c 	if (mdev->interval > 0 && atomic_read(&mdev->busy) == 0 &&
atomic_read       436 drivers/sh/maple/maple.c 			if (atomic_read(&mdev->busy) == 0) {
atomic_read       125 drivers/soc/fsl/qbman/qman_test_stash.c 		while (!atomic_read(&bstrap.started))
atomic_read       353 drivers/soc/ti/knav_dma.c 		if (atomic_read(&chan->ref_count))
atomic_read       363 drivers/soc/ti/knav_dma.c 		if (atomic_read(&dma->ref_count)) {
atomic_read       489 drivers/soc/ti/knav_dma.c 	if (atomic_read(&chan->ref_count) >= 1) {
atomic_read        64 drivers/soc/ti/knav_qmss_acc.c 	if (!enabled || atomic_read(&kq->desc_count) <= 0)
atomic_read       110 drivers/soc/ti/knav_qmss_acc.c 	if (atomic_read(&acc->retrigger_count)) {
atomic_read        90 drivers/soc/ti/knav_qmss_queue.c 		if (atomic_read(&qh->notifier_enabled) <= 0)
atomic_read       422 drivers/soc/ti/knav_qmss_queue.c 		atomic_read(&inst->desc_count);
atomic_read       562 drivers/soc/ti/knav_qmss_queue.c 	while (atomic_read(&qh->notifier_enabled) > 0)
atomic_read      1176 drivers/spi/spi-pxa2xx.c 	if (atomic_read(&drv_data->dma_running))
atomic_read       260 drivers/spi/spi-rockchip.c 	if (atomic_read(&rs->state) & TXDMA)
atomic_read       263 drivers/spi/spi-rockchip.c 	if (atomic_read(&rs->state) & RXDMA)
atomic_read       763 drivers/staging/android/ashmem.c 	wait_event(ashmem_shrink_wait, !atomic_read(&ashmem_shrink_inflight));
atomic_read       456 drivers/staging/android/vsoc.c 		if (atomic_read(address) != arg->value) {
atomic_read       596 drivers/staging/android/vsoc.c 			 (atomic_read(reg_data->incoming_signalled) != 0));
atomic_read       556 drivers/staging/fieldbus/anybuss/host.c 	if (atomic_read(&cd->ind_ab) & pd->flags) {
atomic_read       574 drivers/staging/fieldbus/anybuss/host.c 	if (!(atomic_read(&cd->ind_ab) & pd->flags)) {
atomic_read       704 drivers/staging/fieldbus/anybuss/host.c 	if (((atomic_read(&cd->ind_ab) ^ ind_ap) & IND_AX_MOUT) == 0) {
atomic_read       729 drivers/staging/fieldbus/anybuss/host.c 	if ((atomic_read(&cd->ind_ab) ^ ind_ap) & IND_AX_MIN) {
atomic_read       908 drivers/staging/fieldbus/anybuss/host.c 	ind_ap |= atomic_read(&cd->ind_ab) & IND_AX_EVNT;
atomic_read       925 drivers/staging/fieldbus/anybuss/host.c 	if (!((atomic_read(&cd->ind_ab) ^ ind_ap) & IND_AX_EVNT))
atomic_read       988 drivers/staging/fieldbus/anybuss/host.c 		ind_ab = atomic_read(&cd->ind_ab);
atomic_read       992 drivers/staging/fieldbus/anybuss/host.c 				   (atomic_read(&cd->ind_ab) != ind_ab) ||
atomic_read       420 drivers/staging/greybus/loopback.c 		   !atomic_read(&gb->outstanding_operations));
atomic_read       829 drivers/staging/greybus/loopback.c 				 (atomic_read(&gb->outstanding_operations) <
atomic_read       775 drivers/staging/isdn/gigaset/bas-gigaset.c 			atomic_read(&bcs->hw.bas->corrbytes));
atomic_read      1098 drivers/staging/isdn/gigaset/bas-gigaset.c 		corrbytes = atomic_read(&ubc->corrbytes);
atomic_read       168 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->sleepstatus.status) == 0) {
atomic_read       179 drivers/staging/ks7010/ks7010_sdio.c 	priv->sleep_mode = atomic_read(&priv->sleepstatus.status);
atomic_read       189 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->sleepstatus.status) == 1) {
atomic_read       201 drivers/staging/ks7010/ks7010_sdio.c 	priv->sleep_mode = atomic_read(&priv->sleepstatus.status);
atomic_read       208 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) {
atomic_read       235 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->psstatus.status) == PS_SNOOZE)
atomic_read       244 drivers/staging/ks7010/ks7010_sdio.c 		   atomic_read(&priv->psstatus.status),
atomic_read       245 drivers/staging/ks7010/ks7010_sdio.c 		   atomic_read(&priv->psstatus.confirm_wait),
atomic_read       246 drivers/staging/ks7010/ks7010_sdio.c 		   atomic_read(&priv->psstatus.snooze_guard),
atomic_read       249 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->psstatus.confirm_wait) ||
atomic_read       250 drivers/staging/ks7010/ks7010_sdio.c 	    atomic_read(&priv->psstatus.snooze_guard) ||
atomic_read       356 drivers/staging/ks7010/ks7010_sdio.c 	    atomic_read(&priv->psstatus.status) == PS_SNOOZE)
atomic_read       466 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->psstatus.confirm_wait) && is_hif_conf(event)) {
atomic_read       501 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) {
atomic_read       510 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->sleepstatus.doze_request) == 1) {
atomic_read       515 drivers/staging/ks7010/ks7010_sdio.c 	if (atomic_read(&priv->sleepstatus.wakeup_request) == 1) {
atomic_read       524 drivers/staging/ks7010/ks7010_sdio.c 			   atomic_read(&priv->psstatus.status));
atomic_read       565 drivers/staging/ks7010/ks7010_sdio.c 	    atomic_read(&priv->psstatus.status) == PS_SNOOZE) {
atomic_read       572 drivers/staging/ks7010/ks7010_sdio.c 			if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) {
atomic_read       592 drivers/staging/ks7010/ks7010_sdio.c 			if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) {
atomic_read        74 drivers/staging/ks7010/ks_hostif.c 	ps_status = atomic_read(&priv->psstatus.status);
atomic_read        89 drivers/staging/ks7010/ks_hostif.c 	if (atomic_read(&priv->sme_task.count) > 0)
atomic_read       666 drivers/staging/ks7010/ks_hostif.c 		if (atomic_read(&priv->psstatus.snooze_guard))
atomic_read       670 drivers/staging/ks7010/ks_hostif.c 		if (atomic_read(&priv->psstatus.snooze_guard))
atomic_read      1098 drivers/staging/ks7010/ks_hostif.c 	if (atomic_read(&priv->psstatus.status) == PS_SNOOZE) {
atomic_read      1239 drivers/staging/ks7010/ks_hostif.c 	if (atomic_read(&priv->psstatus.status) > PS_ACTIVE_SET)
atomic_read      1591 drivers/staging/ks7010/ks_hostif.c 	priv->dev_data[atomic_read(&priv->rec_count)] = p;
atomic_read      1592 drivers/staging/ks7010/ks_hostif.c 	priv->dev_size[atomic_read(&priv->rec_count)] = size;
atomic_read      1594 drivers/staging/ks7010/ks_hostif.c 	if (atomic_read(&priv->event_count) != DEVICE_STOCK_COUNT) {
atomic_read      1599 drivers/staging/ks7010/ks_hostif.c 	if (atomic_read(&priv->rec_count) == DEVICE_STOCK_COUNT)
atomic_read        69 drivers/staging/ks7010/ks_wlan_net.c 	if (atomic_read(&update_phyinfo))
atomic_read       640 drivers/staging/ks7010/ks_wlan_net.c 		   in_interrupt(), atomic_read(&update_phyinfo));
atomic_read       646 drivers/staging/ks7010/ks_wlan_net.c 	if (!atomic_read(&update_phyinfo))
atomic_read      1746 drivers/staging/ks7010/ks_wlan_net.c 	if (!atomic_read(&update_phyinfo))
atomic_read        99 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (atomic_read(&imgu_sd->running_mode) == IPU3_RUNNING_MODE_VIDEO)
atomic_read       671 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (atomic_read(&imgu_sd->running_mode) == IPU3_RUNNING_MODE_VIDEO)
atomic_read       572 drivers/staging/media/ipu3/ipu3.c 	if (!atomic_read(&imgu->qbuf_barrier))
atomic_read       196 drivers/staging/media/meson/vdec/esparser.c 	    atomic_read(&sess->esparser_queued_bufs) >= num_dst_bufs)
atomic_read       276 drivers/staging/media/meson/vdec/vdec_helpers.c 	    atomic_read(&sess->esparser_queued_bufs) <= 2) {
atomic_read       284 drivers/staging/media/meson/vdec/vdec_helpers.c 			atomic_read(&sess->esparser_queued_bufs));
atomic_read       657 drivers/staging/media/omap4iss/iss.c 	if (!wait_event_timeout(*wait, !atomic_read(stopping),
atomic_read       459 drivers/staging/media/omap4iss/iss_video.c 		buf->vb.sequence = atomic_read(&pipe->frame_number);
atomic_read       435 drivers/staging/octeon/ethernet-rx.c 	if (!atomic_read(&oct_rx_ready))
atomic_read       120 drivers/staging/octeon/ethernet.c 	if (!atomic_read(&cvm_oct_poll_queue_stopping))
atomic_read       136 drivers/staging/octeon/ethernet.c 	if (!atomic_read(&cvm_oct_poll_queue_stopping))
atomic_read      1672 drivers/staging/qlge/qlge_dbg.c 	pr_err("tx_ring->tx_count = %d\n", atomic_read(&tx_ring->tx_count));
atomic_read       554 drivers/staging/qlge/qlge_ethtool.c 	return atomic_read(&qdev->lb_count) ? -EIO : 0;
atomic_read       673 drivers/staging/qlge/qlge_main.c 	if (!atomic_read(&ctx->irq_cnt)) {
atomic_read      2239 drivers/staging/qlge/qlge_main.c 		if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4)))
atomic_read      2504 drivers/staging/qlge/qlge_main.c 	if (atomic_read(&qdev->intr_context[0].irq_cnt)) {
atomic_read      2648 drivers/staging/qlge/qlge_main.c 	if (unlikely(atomic_read(&tx_ring->tx_count) < 2)) {
atomic_read      2704 drivers/staging/qlge/qlge_main.c 	if (unlikely(atomic_read(&tx_ring->tx_count) < 2)) {
atomic_read      2706 drivers/staging/qlge/qlge_main.c 		if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4)))
atomic_read       721 drivers/staging/ralink-gdma/ralink-gdma.c 		if (atomic_read(&dma_dev->cnt) >= 2) {
atomic_read        44 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c 		if (atomic_read(&entry->refcnt) != 0 && !force)
atomic_read       583 drivers/staging/rtl8192u/r8192U_core.c 		   atomic_read(&(priv->tx_pending[VI_PRIORITY])),
atomic_read       584 drivers/staging/rtl8192u/r8192U_core.c 		   atomic_read(&(priv->tx_pending[VO_PRIORITY])),
atomic_read       585 drivers/staging/rtl8192u/r8192U_core.c 		   atomic_read(&(priv->tx_pending[BE_PRIORITY])),
atomic_read       586 drivers/staging/rtl8192u/r8192U_core.c 		   atomic_read(&(priv->tx_pending[BK_PRIORITY])),
atomic_read       652 drivers/staging/rtl8192u/r8192U_core.c 	int used = atomic_read(&priv->tx_pending[queue_index]);
atomic_read      1429 drivers/staging/rtl8192u/r8192U_core.c 	pend = atomic_read(&priv->tx_pending[tcb_desc->queue_index]);
atomic_read      1580 drivers/staging/rtl8192u/r8192U_core.c 					 atomic_read(&priv->tx_pending[tcb_desc->queue_index]),
atomic_read      1591 drivers/staging/rtl8192u/r8192U_core.c 		 atomic_read(&priv->tx_pending[tcb_desc->queue_index]),
atomic_read      1594 drivers/staging/rtl8192u/r819xU_phy.c 		 atomic_read(&priv->ieee80211->atm_swbw));
atomic_read       320 drivers/staging/rtl8723bs/core/rtw_cmd.c 		|| atomic_read(&(pcmdpriv->cmdthd_running)) == false	/* com_thread not running */
atomic_read       393 drivers/staging/rtl8723bs/core/rtw_cmd.c 		atomic_read(&(adapter->cmdpriv.cmdthd_running)) == true &&
atomic_read      1923 drivers/staging/rtl8723bs/core/rtw_mlme.c 	return (atomic_read(&mlmepriv->set_scan_deny) != 0) ? true : false;
atomic_read      6618 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 	if (evt_seq != (atomic_read(&pevt_priv->event_seq) & 0x7f)) {
atomic_read      6621 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 			  (atomic_read(&pevt_priv->event_seq) & 0x7f)));
atomic_read       639 drivers/staging/rtl8723bs/include/drv_types.h 	int	df = atomic_read(&adapter_to_dvobj(padapter)->disable_func);
atomic_read       646 drivers/staging/rtl8723bs/include/drv_types.h 	int	df = atomic_read(&adapter_to_dvobj(padapter)->disable_func);
atomic_read       651 drivers/staging/rtl8723bs/include/drv_types.h #define RTW_IS_FUNC_DISABLED(padapter, func_bit) (atomic_read(&adapter_to_dvobj(padapter)->disable_func) & (func_bit))
atomic_read      1173 drivers/staging/rtl8723bs/os_dep/os_intfs.c 		while (atomic_read(&(pcmdpriv->cmdthd_running)) == true) {
atomic_read       377 drivers/staging/unisys/visorhba/visorhba_main.c 	if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT)
atomic_read       404 drivers/staging/unisys/visorhba/visorhba_main.c 	if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT)
atomic_read       432 drivers/staging/unisys/visorhba/visorhba_main.c 		if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT)
atomic_read       806 drivers/staging/unisys/visorhba/visorhba_main.c 	if (atomic_read(&vdisk->error_count) < VISORHBA_ERROR_COUNT) {
atomic_read       884 drivers/staging/unisys/visorhba/visorhba_main.c 		if (atomic_read(&vdisk->ios_threshold) > 0) {
atomic_read       886 drivers/staging/unisys/visorhba/visorhba_main.c 			if (atomic_read(&vdisk->ios_threshold) == 0)
atomic_read       978 drivers/staging/unisys/visorhba/visorhba_main.c 			devdata->rsp_queue, (atomic_read(
atomic_read       559 drivers/staging/unisys/visornic/visornic_main.c 	if (atomic_read(&devdata->usage) > 1) {
atomic_read       565 drivers/staging/unisys/visornic/visornic_main.c 			if (atomic_read(&devdata->usage))
atomic_read      1514 drivers/staging/unisys/visornic/visornic_main.c 				     atomic_read(&devdata->num_rcvbuf_in_iovm));
atomic_read       769 drivers/staging/uwb/uwb.h 	unsigned samples = atomic_read(&stats->samples);
atomic_read       799 drivers/staging/uwb/uwb.h 	int samples = atomic_read(&stats->samples);
atomic_read        71 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c 	pos = atomic_read(&alsa_stream->pos);
atomic_read       312 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c 		atomic_read(&alsa_stream->pos));
atomic_read       625 drivers/staging/vc04_services/bcm2835-camera/bcm2835-camera.c 	while (atomic_read(&port->buffers_with_vpu)) {
atomic_read       628 drivers/staging/vc04_services/bcm2835-camera/bcm2835-camera.c 			 __func__, atomic_read(&port->buffers_with_vpu));
atomic_read       634 drivers/staging/vc04_services/bcm2835-camera/bcm2835-camera.c 				 atomic_read(&port->buffers_with_vpu));
atomic_read       513 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c 			value = atomic_read(&service->poll_flags);
atomic_read       518 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c 			value = atomic_read(&state->poll_services[
atomic_read       176 drivers/staging/wusbcore/wa-nep.c 	if (atomic_read(&wa->notifs_queued) > 200) {
atomic_read      1462 drivers/staging/wusbcore/wa-xfer.c 	while (atomic_read(&rpipe->segs_available) > 0
atomic_read      1481 drivers/staging/wusbcore/wa-xfer.c 			atomic_read(&rpipe->segs_available), result);
atomic_read      1506 drivers/staging/wusbcore/wa-xfer.c 		&& (atomic_read(&rpipe->segs_available) ==
atomic_read      1558 drivers/staging/wusbcore/wa-xfer.c 	BUG_ON(atomic_read(&rpipe->segs_available) > maxrequests);
atomic_read      1564 drivers/staging/wusbcore/wa-xfer.c 		available = atomic_read(&rpipe->segs_available);
atomic_read      1604 drivers/staging/wusbcore/wa-xfer.c 		&& (atomic_read(&rpipe->segs_available) ==
atomic_read      2746 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      2804 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      2959 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3002 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3083 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3223 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3303 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3511 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn = cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3552 drivers/target/iscsi/iscsi_target.c 	hdr->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read      3685 drivers/target/iscsi/iscsi_target.c 			if (atomic_read(&conn->check_immediate_queue))
atomic_read      3762 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&conn->check_immediate_queue))
atomic_read      4162 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&conn->connection_recovery)) {
atomic_read      4176 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&conn->conn_logout_remove)) {
atomic_read      4197 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&conn->connection_recovery))
atomic_read      4208 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&conn->sleep_on_conn_wait_comp)) {
atomic_read      4221 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&conn->connection_wait_rcfr)) {
atomic_read      4258 drivers/target/iscsi/iscsi_target.c 		" %s\n", atomic_read(&sess->nconn),
atomic_read      4265 drivers/target/iscsi/iscsi_target.c 	     !atomic_read(&sess->session_logout))
atomic_read      4274 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&sess->nconn)) {
atomic_read      4275 drivers/target/iscsi/iscsi_target.c 		if (!atomic_read(&sess->session_reinstatement) &&
atomic_read      4276 drivers/target/iscsi/iscsi_target.c 		    !atomic_read(&sess->session_fall_back_to_erl0)) {
atomic_read      4280 drivers/target/iscsi/iscsi_target.c 		if (!atomic_read(&sess->session_stop_active)) {
atomic_read      4303 drivers/target/iscsi/iscsi_target.c 	if (!atomic_read(&sess->session_reinstatement) &&
atomic_read      4304 drivers/target/iscsi/iscsi_target.c 	     atomic_read(&sess->session_fall_back_to_erl0)) {
atomic_read      4310 drivers/target/iscsi/iscsi_target.c 	} else if (atomic_read(&sess->session_logout)) {
atomic_read      4314 drivers/target/iscsi/iscsi_target.c 		if (atomic_read(&sess->session_close)) {
atomic_read      4327 drivers/target/iscsi/iscsi_target.c 		if (!atomic_read(&sess->session_continuation))
atomic_read      4330 drivers/target/iscsi/iscsi_target.c 		if (atomic_read(&sess->session_close)) {
atomic_read      4351 drivers/target/iscsi/iscsi_target.c 	if (atomic_read(&sess->nconn)) {
atomic_read      4353 drivers/target/iscsi/iscsi_target.c 			" to %s\n", atomic_read(&sess->nconn),
atomic_read      4584 drivers/target/iscsi/iscsi_target.c 	u16 conn_count = atomic_read(&sess->nconn);
atomic_read      4618 drivers/target/iscsi/iscsi_target.c 	if (session_sleep && atomic_read(&sess->nconn)) {
atomic_read      4644 drivers/target/iscsi/iscsi_target.c 		if (atomic_read(&sess->session_fall_back_to_erl0) ||
atomic_read      4645 drivers/target/iscsi/iscsi_target.c 		    atomic_read(&sess->session_logout) ||
atomic_read      4646 drivers/target/iscsi/iscsi_target.c 		    atomic_read(&sess->session_close) ||
atomic_read       554 drivers/target/iscsi/iscsi_target_configfs.c 		max_cmd_sn = (u32) atomic_read(&sess->max_cmd_sn);
atomic_read      1477 drivers/target/iscsi/iscsi_target_configfs.c 	if (atomic_read(&sess->session_fall_back_to_erl0) ||
atomic_read      1478 drivers/target/iscsi/iscsi_target_configfs.c 	    atomic_read(&sess->session_logout) ||
atomic_read      1479 drivers/target/iscsi/iscsi_target_configfs.c 	    atomic_read(&sess->session_close) ||
atomic_read       755 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&sess->session_reinstatement)) {
atomic_read       825 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&conn->connection_exit)) {
atomic_read       830 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&conn->transport_failed)) {
atomic_read       849 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&conn->connection_exit)) {
atomic_read       854 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&conn->transport_failed)) {
atomic_read       859 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&conn->connection_reinstatement)) {
atomic_read       896 drivers/target/iscsi/iscsi_target_erl0.c 	    !atomic_read(&sess->session_reinstatement) &&
atomic_read       897 drivers/target/iscsi/iscsi_target_erl0.c 	    !atomic_read(&sess->session_fall_back_to_erl0))
atomic_read       912 drivers/target/iscsi/iscsi_target_erl0.c 	if (atomic_read(&conn->connection_exit)) {
atomic_read       157 drivers/target/iscsi/iscsi_target_login.c 		if (atomic_read(&sess_p->session_fall_back_to_erl0) ||
atomic_read       158 drivers/target/iscsi/iscsi_target_login.c 		    atomic_read(&sess_p->session_logout) ||
atomic_read       159 drivers/target/iscsi/iscsi_target_login.c 		    atomic_read(&sess_p->session_close) ||
atomic_read       487 drivers/target/iscsi/iscsi_target_login.c 		if (atomic_read(&sess_p->session_fall_back_to_erl0) ||
atomic_read       488 drivers/target/iscsi/iscsi_target_login.c 		    atomic_read(&sess_p->session_logout) ||
atomic_read       489 drivers/target/iscsi/iscsi_target_login.c 		    atomic_read(&sess_p->session_close) ||
atomic_read       602 drivers/target/iscsi/iscsi_target_login.c 	if ((atomic_read(&sess->nconn) + 1) > sess->sess_ops->MaxConnections) {
atomic_read       718 drivers/target/iscsi/iscsi_target_login.c 			" from node: %s\n", atomic_read(&sess->nconn),
atomic_read       766 drivers/target/iscsi/iscsi_target_login.c 		" %s\n", atomic_read(&sess->nconn),
atomic_read       340 drivers/target/iscsi/iscsi_target_nego.c 	login_rsp->max_cmdsn		= cpu_to_be32((u32) atomic_read(&conn->sess->max_cmd_sn));
atomic_read        45 drivers/target/iscsi/iscsi_target_tmr.c 			iscsi_sna_lte(be32_to_cpu(hdr->refcmdsn), (u32) atomic_read(&conn->sess->max_cmd_sn))) ?
atomic_read       269 drivers/target/iscsi/iscsi_target_util.c 	max_cmdsn = atomic_read(&sess->max_cmd_sn);
atomic_read       571 drivers/target/iscsi/iscsi_target_util.c 	if (!atomic_read(&cmd->immed_queue_count)) {
atomic_read       586 drivers/target/iscsi/iscsi_target_util.c 	if (atomic_read(&cmd->immed_queue_count)) {
atomic_read       589 drivers/target/iscsi/iscsi_target_util.c 			atomic_read(&cmd->immed_queue_count));
atomic_read       647 drivers/target/iscsi/iscsi_target_util.c 	if (!atomic_read(&cmd->response_queue_count)) {
atomic_read       663 drivers/target/iscsi/iscsi_target_util.c 	if (atomic_read(&cmd->response_queue_count)) {
atomic_read       666 drivers/target/iscsi/iscsi_target_util.c 			atomic_read(&cmd->response_queue_count));
atomic_read       771 drivers/target/loopback/tcm_loop.c 	if (atomic_read(&tpg->tl_tpg_port_count)) {
atomic_read       773 drivers/target/loopback/tcm_loop.c 		       atomic_read(&tpg->tl_tpg_port_count));
atomic_read       687 drivers/target/target_core_alua.c 	if (atomic_read(&lun->lun_tg_pt_secondary_offline)) {
atomic_read      1213 drivers/target/target_core_alua.c 			atomic_read(&lun->lun_tg_pt_secondary_offline),
atomic_read      1506 drivers/target/target_core_alua.c 	while (atomic_read(&lu_gp->lu_gp_ref_cnt))
atomic_read      1554 drivers/target/target_core_alua.c 	while (atomic_read(&lu_gp_mem->lu_gp_mem_ref_cnt))
atomic_read      1762 drivers/target/target_core_alua.c 	while (atomic_read(&tg_pt_gp->tg_pt_gp_ref_cnt))
atomic_read      1899 drivers/target/target_core_alua.c 			atomic_read(&lun->lun_tg_pt_secondary_offline) ?
atomic_read      2164 drivers/target/target_core_alua.c 		atomic_read(&lun->lun_tg_pt_secondary_offline));
atomic_read       484 drivers/target/target_core_configfs.c 			BUG_ON(atomic_read(&t->tf_access_cnt));
atomic_read       278 drivers/target/target_core_iblock.c 	if (atomic_read(&ibr->ib_bio_err_cnt))
atomic_read      1321 drivers/target/target_core_pr.c 	while (atomic_read(&pr_reg->pr_res_holders) != 0) {
atomic_read       287 drivers/target/target_core_tpg.c 	while (atomic_read(&nacl->acl_pr_ref_count) != 0)
atomic_read       526 drivers/target/target_core_tpg.c 	while (atomic_read(&se_tpg->tpg_pr_ref_count) != 0)
atomic_read      2025 drivers/target/target_core_transport.c 		if (!atomic_read(&dev->simple_cmds))
atomic_read      2036 drivers/target/target_core_transport.c 	if (atomic_read(&dev->dev_ordered_sync) == 0)
atomic_read      2280 drivers/target/target_core_transport.c 	if (atomic_read(&cmd->se_dev->dev_qf_count) != 0)
atomic_read       224 drivers/target/target_core_user.c 	if (atomic_read(&global_db_count) > tcmu_global_max_blocks)
atomic_read      1271 drivers/target/target_core_user.c 			if (atomic_read(&global_db_count) >
atomic_read      2641 drivers/target/target_core_user.c 	if (atomic_read(&global_db_count) <= tcmu_global_max_blocks)
atomic_read      2694 drivers/target/target_core_user.c 	if (atomic_read(&global_db_count) > tcmu_global_max_blocks)
atomic_read       312 drivers/thermal/intel/intel_powerclamp.c 		atomic_read(&idle_wakeup_counter) >
atomic_read       357 drivers/thermal/intel/intel_powerclamp.c 	reduce_irq = atomic_read(&idle_wakeup_counter) >=
atomic_read       479 drivers/thermal/thermal_core.c 	if (atomic_read(&in_suspend))
atomic_read       493 drivers/tty/mips_ejtag_fdc.c 					 atomic_read(&priv->xmit_total) ||
atomic_read       596 drivers/tty/rocket.c 	if (atomic_read(&rp_num_ports_open))
atomic_read       904 drivers/tty/rocket.c 				atomic_read(&rp_num_ports_open));
atomic_read      1035 drivers/tty/rocket.c 			atomic_read(&rp_num_ports_open));
atomic_read       293 drivers/tty/serial/atmel_serial.c 	if (!atomic_read(&atmel_port->tasklet_shutdown))
atomic_read       577 drivers/tty/serial/atmel_serial.c 		if (!atomic_read(&atmel_port->tasklet_shutdown))
atomic_read      1282 drivers/tty/serial/atmel_serial.c 	if (!atomic_read(&atmel_port->tasklet_shutdown)) {
atomic_read       138 drivers/tty/serial/kgdb_nmi.c 	if (atomic_read(&kgdb_nmi_num_readers)) {
atomic_read       199 drivers/tty/serial/kgdb_nmi.c 	if (likely(!atomic_read(&kgdb_nmi_num_readers) ||
atomic_read      2952 drivers/tty/serial/serial_core.c 	wait_event(state->remove_wait, !atomic_read(&state->refcount));
atomic_read        93 drivers/tty/tty_buffer.c 	int space = port->buf.mem_limit - atomic_read(&port->buf.mem_used);
atomic_read       173 drivers/tty/tty_buffer.c 	if (atomic_read(&port->buf.mem_used) > port->buf.mem_limit)
atomic_read       513 drivers/tty/tty_buffer.c 		if (atomic_read(&buf->priority))
atomic_read       259 drivers/uio/uio.c 	return sprintf(buf, "%u\n", (unsigned int)atomic_read(&idev->event));
atomic_read       487 drivers/uio/uio.c 	listener->event_count = atomic_read(&idev->event);
atomic_read       558 drivers/uio/uio.c 	if (listener->event_count != atomic_read(&idev->event))
atomic_read       588 drivers/uio/uio.c 		event_count = atomic_read(&idev->event);
atomic_read       745 drivers/usb/atm/usbatm.c 			       atomic_read(&atm_dev->stats.aal5.tx),
atomic_read       746 drivers/usb/atm/usbatm.c 			       atomic_read(&atm_dev->stats.aal5.tx_err),
atomic_read       747 drivers/usb/atm/usbatm.c 			       atomic_read(&atm_dev->stats.aal5.rx),
atomic_read       748 drivers/usb/atm/usbatm.c 			       atomic_read(&atm_dev->stats.aal5.rx_err),
atomic_read       749 drivers/usb/atm/usbatm.c 			       atomic_read(&atm_dev->stats.aal5.rx_drop));
atomic_read       532 drivers/usb/class/usbtmc.c 			atomic_read(&data->iin_data_valid) != 0,
atomic_read       593 drivers/usb/class/usbtmc.c 			atomic_read(&file_data->srq_asserted) != 0 ||
atomic_read       594 drivers/usb/class/usbtmc.c 			atomic_read(&file_data->closing),
atomic_read       600 drivers/usb/class/usbtmc.c 	if (atomic_read(&file_data->closing) || data->zombie)
atomic_read      2188 drivers/usb/class/usbtmc.c 	if (atomic_read(&file_data->srq_asserted))
atomic_read       632 drivers/usb/core/devices.c 	event_count = atomic_read(&device_event.count);
atomic_read      1573 drivers/usb/core/driver.c 			__func__, atomic_read(&udev->dev.power.usage_count),
atomic_read      1605 drivers/usb/core/driver.c 			__func__, atomic_read(&udev->dev.power.usage_count),
atomic_read      1635 drivers/usb/core/driver.c 			__func__, atomic_read(&intf->dev.power.usage_count),
atomic_read      1663 drivers/usb/core/driver.c 			__func__, atomic_read(&intf->dev.power.usage_count),
atomic_read      1713 drivers/usb/core/driver.c 			__func__, atomic_read(&intf->dev.power.usage_count),
atomic_read      1746 drivers/usb/core/driver.c 			__func__, atomic_read(&intf->dev.power.usage_count),
atomic_read      1796 drivers/usb/core/driver.c 			if (atomic_read(&intf->dev.power.usage_count) > 0)
atomic_read      1151 drivers/usb/core/hcd.c 	if (unlikely(atomic_read(&urb->reject))) {
atomic_read      1564 drivers/usb/core/hcd.c 		if (atomic_read(&urb->reject))
atomic_read      1613 drivers/usb/core/hcd.c 	if (atomic_read(&urb->use_count) > 0) {
atomic_read      1658 drivers/usb/core/hcd.c 	if (unlikely(atomic_read(&urb->reject)))
atomic_read       297 drivers/usb/core/sysfs.c 	return sprintf(buf, "%d\n", atomic_read(&udev->urbnum));
atomic_read       144 drivers/usb/core/urb.c 	return atomic_read(&anchor->suspend_wakeups) == 0 &&
atomic_read       696 drivers/usb/core/urb.c 	wait_event(usb_kill_urb_queue, atomic_read(&urb->use_count) == 0);
atomic_read       740 drivers/usb/core/urb.c 	wait_event(usb_kill_urb_queue, atomic_read(&urb->use_count) == 0);
atomic_read       384 drivers/usb/gadget/function/f_ecm.c 	if (atomic_read(&ecm->notify_count))
atomic_read       913 drivers/usb/gadget/function/f_ecm.c 	if (atomic_read(&ecm->notify_count)) {
atomic_read      1749 drivers/usb/gadget/function/f_fs.c 	if (atomic_read(&ffs->opened) < 0) {
atomic_read       551 drivers/usb/gadget/function/f_ncm.c 	if (atomic_read(&ncm->notify_count))
atomic_read      1656 drivers/usb/gadget/function/f_ncm.c 	if (atomic_read(&ncm->notify_count)) {
atomic_read       439 drivers/usb/gadget/function/f_tcm.c 		luns = atomic_read(&fu->tpg->tpg_port_count);
atomic_read      1615 drivers/usb/gadget/function/f_tcm.c 	if (atomic_read(&tpg->tpg_port_count)) {
atomic_read      1618 drivers/usb/gadget/function/f_tcm.c 		pr_err(MSG, atomic_read(&tpg->tpg_port_count));
atomic_read      2472 drivers/usb/gadget/udc/lpc32xx_udc.c 		if (atomic_read(&udc->enabled_ep_cnt))
atomic_read      2474 drivers/usb/gadget/udc/lpc32xx_udc.c 				 (atomic_read(&udc->enabled_ep_cnt) == 0));
atomic_read      2945 drivers/usb/gadget/udc/lpc32xx_udc.c 		if (atomic_read(&udc->enabled_ep_cnt))
atomic_read      2947 drivers/usb/gadget/udc/lpc32xx_udc.c 				(atomic_read(&udc->enabled_ep_cnt) == 0));
atomic_read       487 drivers/usb/host/isp116x-hcd.c 	if (atomic_read(&isp116x->atl_finishing))
atomic_read       749 drivers/usb/host/isp1362-hcd.c 	if (atomic_read(&epq->finishing)) {
atomic_read       801 drivers/usb/host/isp1362-hcd.c 	if (atomic_read(&epq->finishing)) {
atomic_read       868 drivers/usb/host/isp1362-hcd.c 	if (atomic_read(&epq->finishing)) {
atomic_read       178 drivers/usb/misc/iowarrior.c 	intr_idx = atomic_read(&dev->intr_idx);
atomic_read       181 drivers/usb/misc/iowarrior.c 	read_idx = atomic_read(&dev->read_idx);
atomic_read       268 drivers/usb/misc/iowarrior.c 	read_idx = atomic_read(&dev->read_idx);
atomic_read       269 drivers/usb/misc/iowarrior.c 	intr_idx = atomic_read(&dev->intr_idx);
atomic_read       332 drivers/usb/misc/iowarrior.c 	} while (atomic_read(&dev->overflow_flag));
atomic_read       393 drivers/usb/misc/iowarrior.c 		if (atomic_read(&dev->write_busy) == MAX_WRITES_IN_FLIGHT) {
atomic_read       400 drivers/usb/misc/iowarrior.c 								  (!dev->present || (atomic_read (&dev-> write_busy) < MAX_WRITES_IN_FLIGHT)));
atomic_read       448 drivers/usb/misc/iowarrior.c 				retval, atomic_read(&dev->write_busy));
atomic_read       701 drivers/usb/misc/iowarrior.c 	if (atomic_read(&dev->write_busy) < MAX_WRITES_IN_FLIGHT)
atomic_read       853 drivers/usb/serial/io_edgeport.c 		__func__, urb, atomic_read(&CmdUrbs));
atomic_read      2260 drivers/usb/serial/io_edgeport.c 		__func__, urb, atomic_read(&CmdUrbs));
atomic_read       566 drivers/usb/serial/mos7720.c 	status = atomic_read(&mos_parport->shadowDSR) & 0xf8;
atomic_read       768 drivers/usb/storage/realtek_cr.c 			     atomic_read(&us->pusb_intf->dev.power.usage_count));
atomic_read       770 drivers/usb/storage/realtek_cr.c 		if (atomic_read(&us->pusb_intf->dev.power.usage_count) > 0) {
atomic_read       777 drivers/usb/storage/realtek_cr.c 				     atomic_read(&us->pusb_intf->dev.power.usage_count));
atomic_read       810 drivers/usb/storage/realtek_cr.c 			     atomic_read(&us->pusb_intf->dev.power.usage_count));
atomic_read       812 drivers/usb/storage/realtek_cr.c 		if (atomic_read(&us->pusb_intf->dev.power.usage_count) <= 0) {
atomic_read        71 drivers/usb/usbip/vhci_rx.c 			atomic_read(&vhci_hcd->seqnum));
atomic_read       701 drivers/vfio/vfio.c 	if (!atomic_read(&group->container_users))
atomic_read       714 drivers/vfio/vfio.c 	if (!atomic_read(&group->container_users))
atomic_read      1355 drivers/vfio/vfio.c 	if (atomic_read(&group->container_users))
atomic_read      1436 drivers/vfio/vfio.c 	if (0 == atomic_read(&group->container_users) ||
atomic_read       260 drivers/vhost/net.c 	wait_event(ubufs->wait, !atomic_read(&ubufs->refcount));
atomic_read       382 drivers/vhost/vsock.c 	val = atomic_read(&vsock->queued_replies);
atomic_read       193 drivers/video/fbdev/arcfb.c 	int count = atomic_read(&par->ref_count);
atomic_read       227 drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c 	if (atomic_read(&rg->map_count)) {
atomic_read       664 drivers/video/fbdev/omap2/omapfb/omapfb-main.c 	WARN_ON(!atomic_read(&ofbi->region->lock_count));
atomic_read       847 drivers/video/fbdev/omap2/omapfb/omapfb-main.c 	WARN_ON(!atomic_read(&ofbi->region->lock_count));
atomic_read       941 drivers/video/fbdev/omap2/omapfb/omapfb-main.c 	WARN_ON(!atomic_read(&ofbi->region->lock_count));
atomic_read      1312 drivers/video/fbdev/omap2/omapfb/omapfb-main.c 	WARN_ON(atomic_read(&rg->map_count));
atomic_read       449 drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c 	if (atomic_read(&rg->map_count)) {
atomic_read       513 drivers/video/fbdev/ps3fb.c 		if (atomic_read(&ps3fb.ext_flip)) {
atomic_read       926 drivers/video/fbdev/ps3fb.c 		    !atomic_read(&ps3fb.ext_flip)) {
atomic_read       860 drivers/video/fbdev/smscufx.c 	if (!atomic_read(&dev->usb_active))
atomic_read       965 drivers/video/fbdev/smscufx.c 	if (!atomic_read(&dev->usb_active))
atomic_read       991 drivers/video/fbdev/smscufx.c 	if (!atomic_read(&dev->usb_active))
atomic_read       288 drivers/video/fbdev/udlfb.c 	if (!atomic_read(&dlfb->usb_active))
atomic_read       619 drivers/video/fbdev/udlfb.c 	if (!atomic_read(&dlfb->usb_active)) {
atomic_read       799 drivers/video/fbdev/udlfb.c 	if (!atomic_read(&dlfb->usb_active))
atomic_read       877 drivers/video/fbdev/udlfb.c 	if (!atomic_read(&dlfb->usb_active))
atomic_read      1430 drivers/video/fbdev/udlfb.c 			atomic_read(&dlfb->bytes_rendered));
atomic_read      1438 drivers/video/fbdev/udlfb.c 			atomic_read(&dlfb->bytes_identical));
atomic_read      1446 drivers/video/fbdev/udlfb.c 			atomic_read(&dlfb->bytes_sent));
atomic_read      1454 drivers/video/fbdev/udlfb.c 			atomic_read(&dlfb->cpu_kcycles_used));
atomic_read      1166 drivers/video/fbdev/uvesafb.c 	int cnt = atomic_read(&par->ref_count);
atomic_read      1187 drivers/video/fbdev/uvesafb.c 	int cnt = atomic_read(&par->ref_count);
atomic_read        75 drivers/w1/slaves/w1_therm.c 		refcnt = atomic_read(THERM_REFCNT(sl->family_data));
atomic_read        73 drivers/w1/w1_family.c 	while (atomic_read(&fent->refcnt)) {
atomic_read        75 drivers/w1/w1_family.c 				fent->fid, atomic_read(&fent->refcnt));
atomic_read       198 drivers/w1/w1_int.c 	while (atomic_read(&dev->refcnt)) {
atomic_read       200 drivers/w1/w1_int.c 				dev->name, atomic_read(&dev->refcnt));
atomic_read       249 drivers/watchdog/pcwd_usb.c 		if (atomic_read(&usb_pcwd->cmd_received))
atomic_read       231 drivers/xen/pvcalls-back.c 	while (atomic_read(&map->io) > 0) {
atomic_read       232 drivers/xen/pvcalls-back.c 		if (atomic_read(&map->release) > 0) {
atomic_read       237 drivers/xen/pvcalls-back.c 		if (atomic_read(&map->read) > 0)
atomic_read       239 drivers/xen/pvcalls-back.c 		if (atomic_read(&map->write) > 0)
atomic_read      1054 drivers/xen/pvcalls-front.c 		while (atomic_read(&map->refcount) > 1)
atomic_read      1062 drivers/xen/pvcalls-front.c 		while (atomic_read(&map->refcount) > 1)
atomic_read      1106 drivers/xen/pvcalls-front.c 	while (atomic_read(&pvcalls_refcount) > 0)
atomic_read       586 drivers/xen/xen-scsiback.c 		atomic_read(&info->nr_unreplied_reqs) == 0);
atomic_read        25 fs/affs/file.c 		 inode->i_ino, atomic_read(&AFFS_I(inode)->i_opencnt));
atomic_read        34 fs/affs/file.c 		 inode->i_ino, atomic_read(&AFFS_I(inode)->i_opencnt));
atomic_read       482 fs/afs/cell.c  	ASSERTCMP(atomic_read(&cell->usage), ==, 0);
atomic_read       691 fs/afs/cell.c  		if (atomic_read(&cell->usage) > 1) {
atomic_read       704 fs/afs/cell.c  		if (atomic_read(&cell->usage) > 1)
atomic_read       738 fs/afs/cell.c  	_leave(" [destruct %d]", atomic_read(&net->cells_outstanding));
atomic_read       775 fs/afs/cell.c  		usage = atomic_read(&cell->usage);
atomic_read       830 fs/afs/cell.c  	_leave(" [%d]", atomic_read(&net->cells_outstanding));
atomic_read       857 fs/afs/cell.c  		       !atomic_read(&net->cells_outstanding));
atomic_read       271 fs/afs/cmservice.c 		trace_afs_server(call->server, atomic_read(&call->server->usage),
atomic_read       194 fs/afs/dynroot.c 	if (!sb || atomic_read(&sb->s_active) == 0)
atomic_read       223 fs/afs/dynroot.c 	if (!sb || atomic_read(&sb->s_active) == 0)
atomic_read        50 fs/afs/proc.c  		   atomic_read(&cell->usage),
atomic_read       221 fs/afs/proc.c  		   atomic_read(&vol->usage), vol->vid,
atomic_read       388 fs/afs/proc.c  		   atomic_read(&server->usage),
atomic_read       578 fs/afs/proc.c  		   atomic_read(&net->n_lookup),
atomic_read       579 fs/afs/proc.c  		   atomic_read(&net->n_reval),
atomic_read       580 fs/afs/proc.c  		   atomic_read(&net->n_inval),
atomic_read       581 fs/afs/proc.c  		   atomic_read(&net->n_relpg));
atomic_read       584 fs/afs/proc.c  		   atomic_read(&net->n_read_dir));
atomic_read       587 fs/afs/proc.c  		   atomic_read(&net->n_dir_cr),
atomic_read       588 fs/afs/proc.c  		   atomic_read(&net->n_dir_rm));
atomic_read       591 fs/afs/proc.c  		   atomic_read(&net->n_fetches),
atomic_read       594 fs/afs/proc.c  		   atomic_read(&net->n_stores),
atomic_read       120 fs/afs/rxrpc.c 	_debug("outstanding %u", atomic_read(&net->nr_outstanding_calls));
atomic_read       122 fs/afs/rxrpc.c 		       !atomic_read(&net->nr_outstanding_calls));
atomic_read       169 fs/afs/rxrpc.c 	int o = atomic_read(&net->nr_outstanding_calls);
atomic_read       207 fs/afs/rxrpc.c 		       atomic_read(&call->net->nr_outstanding_calls),
atomic_read       377 fs/afs/rxrpc.c 	       atomic_read(&call->net->nr_outstanding_calls));
atomic_read       709 fs/afs/rxrpc.c 			       atomic_read(&call->net->nr_outstanding_calls),
atomic_read       295 fs/afs/server.c 	_leave(" = %p{%d}", server, atomic_read(&server->usage));
atomic_read       364 fs/afs/server.c 	trace_afs_server(server, atomic_read(&server->usage),
atomic_read       382 fs/afs/server.c 	trace_afs_server(server, atomic_read(&server->usage),
atomic_read       389 fs/afs/server.c 		       atomic_read(&server->probe_outstanding) == 0);
atomic_read       391 fs/afs/server.c 	trace_afs_server(server, atomic_read(&server->usage),
atomic_read       456 fs/afs/server.c 		int usage = atomic_read(&server->usage);
atomic_read       498 fs/afs/server.c 	_leave(" [%d]", atomic_read(&net->servers_outstanding));
atomic_read       522 fs/afs/server.c 		       !atomic_read(&net->servers_outstanding));
atomic_read       535 fs/afs/server.c 	trace_afs_server(server, atomic_read(&server->usage), afs_server_trace_update);
atomic_read       143 fs/afs/super.c 	if (atomic_read(&afs_count_active_inodes) != 0) {
atomic_read       145 fs/afs/super.c 		       atomic_read(&afs_count_active_inodes));
atomic_read       343 fs/aio.c       			if (!atomic_read(&ctx->dead)) {
atomic_read       929 fs/aio.c       		int old, avail = atomic_read(&ctx->reqs_available);
atomic_read      1262 fs/aio.c       	if (unlikely(atomic_read(&ctx->dead)))
atomic_read       214 fs/autofs/expire.c 			unsigned int ino_count = atomic_read(&ino->count);
atomic_read       382 fs/autofs/expire.c 			ino_count = atomic_read(&ino->count) + 1;
atomic_read       399 fs/autofs/expire.c 			ino_count = atomic_read(&ino->count) + 1;
atomic_read        79 fs/btrfs/async-thread.c 	return atomic_read(&wq->normal->pending) > wq->normal->thresh * 2;
atomic_read       204 fs/btrfs/async-thread.c 	pending = atomic_read(&wq->pending);
atomic_read       319 fs/btrfs/block-group.c 	wait_var_event(&bg->nocow_writers, !atomic_read(&bg->nocow_writers));
atomic_read       356 fs/btrfs/block-group.c 	wait_var_event(&bg->reservations, !atomic_read(&bg->reservations));
atomic_read      1090 fs/btrfs/block-group.c 	remove_em = (atomic_read(&block_group->trimming) == 0);
atomic_read      2604 fs/btrfs/block-group.c 				   atomic_read(&cur_trans->num_writers) == 1);
atomic_read      3145 fs/btrfs/block-group.c 		ASSERT(atomic_read(&block_group->count) == 1);
atomic_read       916 fs/btrfs/compression.c 	if (atomic_read(total_ws) > cpus) {
atomic_read       921 fs/btrfs/compression.c 		if (atomic_read(total_ws) > cpus && !*free_ws)
atomic_read       952 fs/btrfs/compression.c 		if (atomic_read(total_ws) == 0) {
atomic_read      1331 fs/btrfs/delayed-inode.c 		if (atomic_read(&delayed_root->items) <
atomic_read      1398 fs/btrfs/delayed-inode.c 	int val = atomic_read(&delayed_root->items_seq);
atomic_read      1403 fs/btrfs/delayed-inode.c 	if (atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND)
atomic_read      1413 fs/btrfs/delayed-inode.c 	if ((atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND) ||
atomic_read      1417 fs/btrfs/delayed-inode.c 	if (atomic_read(&delayed_root->items) >= BTRFS_DELAYED_WRITEBACK) {
atomic_read      1421 fs/btrfs/delayed-inode.c 		seq = atomic_read(&delayed_root->items_seq);
atomic_read        56 fs/btrfs/delayed-ref.c 		atomic_read(&trans->transaction->delayed_refs.num_entries);
atomic_read       871 fs/btrfs/disk-io.c 	if (atomic_read(&bi->sync_writers))
atomic_read      1006 fs/btrfs/disk-io.c 	BUG_ON(!atomic_read(&eb->refs));
atomic_read      4006 fs/btrfs/disk-io.c 		   (atomic_read(&fs_info->defrag_running) == 0));
atomic_read      4276 fs/btrfs/disk-io.c 	if (atomic_read(&delayed_refs->num_entries) == 0) {
atomic_read      4582 fs/btrfs/disk-io.c 				   atomic_read(&t->num_writers) == 0);
atomic_read      2180 fs/btrfs/extent-tree.c 		count = atomic_read(&delayed_refs->num_entries) * 2;
atomic_read      5751 fs/btrfs/extent-tree.c 	if (atomic_read(&root->will_be_snapshotted))
atomic_read      5759 fs/btrfs/extent-tree.c 	if (atomic_read(&root->will_be_snapshotted)) {
atomic_read      5775 fs/btrfs/extent-tree.c 			       !atomic_read(&root->will_be_snapshotted));
atomic_read        81 fs/btrfs/extent_io.c 		       eb->start, eb->len, atomic_read(&eb->refs), eb->bflags);
atomic_read      4856 fs/btrfs/extent_io.c 	return (atomic_read(&eb->io_pages) ||
atomic_read      5048 fs/btrfs/extent_io.c 	refs = atomic_read(&eb->refs);
atomic_read      5287 fs/btrfs/extent_io.c 	WARN_ON(atomic_read(&eb->refs) == 0);
atomic_read      5326 fs/btrfs/extent_io.c 		refs = atomic_read(&eb->refs);
atomic_read      5337 fs/btrfs/extent_io.c 	if (atomic_read(&eb->refs) == 2 &&
atomic_read      5358 fs/btrfs/extent_io.c 	if (atomic_read(&eb->refs) == 2 && !extent_buffer_under_io(eb) &&
atomic_read      5389 fs/btrfs/extent_io.c 	WARN_ON(atomic_read(&eb->refs) == 0);
atomic_read      5403 fs/btrfs/extent_io.c 	WARN_ON(atomic_read(&eb->refs) == 0);
atomic_read      6111 fs/btrfs/extent_io.c 	if (atomic_read(&eb->refs) != 1 || extent_buffer_under_io(eb)) {
atomic_read      1512 fs/btrfs/inode.c 			if (!freespace_inode && atomic_read(&root->snapshot_force_cow))
atomic_read      2090 fs/btrfs/inode.c 	int async = !atomic_read(&BTRFS_I(inode)->sync_writers);
atomic_read      3507 fs/btrfs/inode.c 			atomic_read(&fs_info->nr_delayed_iputs) == 0);
atomic_read      4485 fs/btrfs/inode.c 			if (atomic_read(&inode->i_count) > 1)
atomic_read      8465 fs/btrfs/inode.c 		async_submit = !atomic_read(&BTRFS_I(inode)->sync_writers);
atomic_read       770 fs/btrfs/ioctl.c 	if (atomic_read(&root->nr_swapfiles)) {
atomic_read      4537 fs/btrfs/ioctl.c 	if (atomic_read(&fs_info->balance_pause_req))
atomic_read      4539 fs/btrfs/ioctl.c 	if (atomic_read(&fs_info->balance_cancel_req))
atomic_read        41 fs/btrfs/locking.c 	WARN_ON(atomic_read(&eb->spinning_readers) == 0);
atomic_read        57 fs/btrfs/locking.c 	BUG_ON(!atomic_read(&eb->read_locks));
atomic_read       210 fs/btrfs/locking.c 	if (eb->blocking_writers || atomic_read(&eb->blocking_readers))
atomic_read       214 fs/btrfs/locking.c 	if (eb->blocking_writers || atomic_read(&eb->blocking_readers)) {
atomic_read       264 fs/btrfs/locking.c 	WARN_ON(atomic_read(&eb->blocking_readers) == 0);
atomic_read       284 fs/btrfs/locking.c 	wait_event(eb->read_lock_wq, atomic_read(&eb->blocking_readers) == 0);
atomic_read       287 fs/btrfs/locking.c 	if (atomic_read(&eb->blocking_readers) || eb->blocking_writers) {
atomic_read       156 fs/btrfs/print-tree.c 		   atomic_read(&eb->refs), eb->write_locks,
atomic_read       157 fs/btrfs/print-tree.c 		   atomic_read(&eb->read_locks),
atomic_read       159 fs/btrfs/print-tree.c 		   atomic_read(&eb->blocking_readers),
atomic_read       161 fs/btrfs/print-tree.c 		   atomic_read(&eb->spinning_readers),
atomic_read       922 fs/btrfs/raid56.c 	if (atomic_read(&rbio->error) > max_errors)
atomic_read      1333 fs/btrfs/raid56.c 	BUG_ON(atomic_read(&rbio->stripes_pending) == 0);
atomic_read      1491 fs/btrfs/raid56.c 	if (atomic_read(&rbio->error) > rbio->bbio->max_errors)
atomic_read      2042 fs/btrfs/raid56.c 	if (atomic_read(&rbio->error) > rbio->bbio->max_errors)
atomic_read      2110 fs/btrfs/raid56.c 		if (atomic_read(&rbio->error) <= rbio->bbio->max_errors) {
atomic_read      2532 fs/btrfs/raid56.c 	if (atomic_read(&rbio->error) > rbio->bbio->max_errors)
atomic_read       783 fs/btrfs/reada.c 			if (atomic_read(&device->reada_in_flight) <
atomic_read       807 fs/btrfs/reada.c 		if (atomic_read(&fs_info->reada_works_cnt) >
atomic_read       843 fs/btrfs/reada.c 			atomic_read(&device->reada_in_flight));
atomic_read       976 fs/btrfs/reada.c 	while (atomic_read(&rc->elems)) {
atomic_read       977 fs/btrfs/reada.c 		if (!atomic_read(&fs_info->reada_works_cnt))
atomic_read       979 fs/btrfs/reada.c 		wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0,
atomic_read       981 fs/btrfs/reada.c 		dump_devs(fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0);
atomic_read       984 fs/btrfs/reada.c 	dump_devs(fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0);
atomic_read       996 fs/btrfs/reada.c 	while (atomic_read(&rc->elems)) {
atomic_read       997 fs/btrfs/reada.c 		if (!atomic_read(&fs_info->reada_works_cnt))
atomic_read       999 fs/btrfs/reada.c 		wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0,
atomic_read       279 fs/btrfs/scrub.c 	while (atomic_read(&fs_info->scrub_pause_req)) {
atomic_read       282 fs/btrfs/scrub.c 		   atomic_read(&fs_info->scrub_pause_req) == 0);
atomic_read      3142 fs/btrfs/scrub.c 		   atomic_read(&sctx->bios_in_flight) == 0);
atomic_read      3182 fs/btrfs/scrub.c 		if (atomic_read(&fs_info->scrub_cancel_req) ||
atomic_read      3183 fs/btrfs/scrub.c 		    atomic_read(&sctx->cancel_req)) {
atomic_read      3190 fs/btrfs/scrub.c 		if (atomic_read(&fs_info->scrub_pause_req)) {
atomic_read      3198 fs/btrfs/scrub.c 				   atomic_read(&sctx->bios_in_flight) == 0);
atomic_read      3649 fs/btrfs/scrub.c 			   atomic_read(&sctx->bios_in_flight) == 0);
atomic_read      3659 fs/btrfs/scrub.c 			   atomic_read(&sctx->workers_pending) == 0);
atomic_read      3740 fs/btrfs/scrub.c 	wait_event(sctx->list_wait, atomic_read(&sctx->bios_in_flight) == 0);
atomic_read      3931 fs/btrfs/scrub.c 	wait_event(sctx->list_wait, atomic_read(&sctx->bios_in_flight) == 0);
atomic_read      3935 fs/btrfs/scrub.c 	wait_event(sctx->list_wait, atomic_read(&sctx->workers_pending) == 0);
atomic_read      3974 fs/btrfs/scrub.c 	while (atomic_read(&fs_info->scrubs_paused) !=
atomic_read      3975 fs/btrfs/scrub.c 	       atomic_read(&fs_info->scrubs_running)) {
atomic_read      3978 fs/btrfs/scrub.c 			   atomic_read(&fs_info->scrubs_paused) ==
atomic_read      3979 fs/btrfs/scrub.c 			   atomic_read(&fs_info->scrubs_running));
atomic_read      3994 fs/btrfs/scrub.c 	if (!atomic_read(&fs_info->scrubs_running)) {
atomic_read      4000 fs/btrfs/scrub.c 	while (atomic_read(&fs_info->scrubs_running)) {
atomic_read      4003 fs/btrfs/scrub.c 			   atomic_read(&fs_info->scrubs_running) == 0);
atomic_read       409 fs/btrfs/space-info.c 		async_pages = atomic_read(&fs_info->async_delalloc_pages);
atomic_read       424 fs/btrfs/space-info.c 			   atomic_read(&fs_info->async_delalloc_pages) <=
atomic_read      1699 fs/btrfs/super.c 			   (atomic_read(&fs_info->defrag_running) == 0));
atomic_read       137 fs/btrfs/transaction.c 	return atomic_read(&trans->num_extwriters);
atomic_read       874 fs/btrfs/transaction.c 	WARN_ON(atomic_read(&cur_trans->num_writers) < 1);
atomic_read      1857 fs/btrfs/transaction.c 			   atomic_read(&cur_trans->num_writers) == 1);
atomic_read      2119 fs/btrfs/transaction.c 		   atomic_read(&cur_trans->num_writers) == 1);
atomic_read      2968 fs/btrfs/tree-log.c 		      atomic_read(&root->log_commit[index])))
atomic_read      2985 fs/btrfs/tree-log.c 		if (!atomic_read(&root->log_writers))
atomic_read      3059 fs/btrfs/tree-log.c 	if (atomic_read(&root->log_commit[index1])) {
atomic_read      3068 fs/btrfs/tree-log.c 	if (atomic_read(&root->log_commit[(index1 + 1) % 2]))
atomic_read      3072 fs/btrfs/tree-log.c 		int batch = atomic_read(&root->log_batch);
atomic_read      3081 fs/btrfs/tree-log.c 		if (batch == atomic_read(&root->log_batch))
atomic_read      3189 fs/btrfs/tree-log.c 	if (atomic_read(&log_root_tree->log_commit[index2])) {
atomic_read      3202 fs/btrfs/tree-log.c 	if (atomic_read(&log_root_tree->log_commit[(index2 + 1) % 2])) {
atomic_read       620 fs/btrfs/volumes.c 		BUG_ON(atomic_read(&cur->__bi_cnt) == 0);
atomic_read      3732 fs/btrfs/volumes.c 		if ((!counting && atomic_read(&fs_info->balance_pause_req)) ||
atomic_read      3733 fs/btrfs/volumes.c 		    atomic_read(&fs_info->balance_cancel_req)) {
atomic_read      3903 fs/btrfs/volumes.c 	return atomic_read(&fs_info->balance_cancel_req) ||
atomic_read      3904 fs/btrfs/volumes.c 		(atomic_read(&fs_info->balance_pause_req) == 0 &&
atomic_read      3905 fs/btrfs/volumes.c 		 atomic_read(&fs_info->balance_cancel_req) == 0);
atomic_read      4088 fs/btrfs/volumes.c 	    atomic_read(&fs_info->balance_pause_req) ||
atomic_read      4089 fs/btrfs/volumes.c 	    atomic_read(&fs_info->balance_cancel_req)) {
atomic_read      4235 fs/btrfs/volumes.c 	if (ret == -ECANCELED && atomic_read(&fs_info->balance_pause_req))
atomic_read      4237 fs/btrfs/volumes.c 	else if (ret == -ECANCELED && atomic_read(&fs_info->balance_cancel_req))
atomic_read      6452 fs/btrfs/volumes.c 		if (atomic_read(&bbio->error) > bbio->max_errors) {
atomic_read      6546 fs/btrfs/volumes.c 		if (atomic_read(&bbio->error) > bbio->max_errors)
atomic_read      7505 fs/btrfs/volumes.c 		stats_cnt = atomic_read(&device->dev_stats_ccnt);
atomic_read       515 fs/btrfs/volumes.h 	return atomic_read(dev->dev_stat_values + index);
atomic_read      1141 fs/buffer.c    	if (atomic_read(&buf->b_count)) {
atomic_read      3170 fs/buffer.c    	WARN_ON(atomic_read(&bh->b_count) < 1);
atomic_read      3214 fs/buffer.c    	return atomic_read(&bh->b_count) |
atomic_read       189 fs/cachefiles/interface.c 	_enter("{OBJ%x,%d}", _object->debug_id, atomic_read(&object->usage));
atomic_read       192 fs/cachefiles/interface.c 	ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000);
atomic_read       274 fs/cachefiles/interface.c 	       object->fscache.debug_id, atomic_read(&object->usage));
atomic_read       280 fs/cachefiles/interface.c 	ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000);
atomic_read       334 fs/cachefiles/interface.c 	       object->fscache.debug_id, atomic_read(&object->usage));
atomic_read       337 fs/cachefiles/interface.c 	ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000);
atomic_read        34 fs/cachefiles/proc.c 		x = atomic_read(&cachefiles_lookup_histogram[index]);
atomic_read        35 fs/cachefiles/proc.c 		y = atomic_read(&cachefiles_mkdir_histogram[index]);
atomic_read        36 fs/cachefiles/proc.c 		z = atomic_read(&cachefiles_create_histogram[index]);
atomic_read       702 fs/cachefiles/rdwr.c 	       object->fscache.debug_id, atomic_read(&op->op.usage),
atomic_read      2764 fs/ceph/caps.c 		flags = atomic_read(&fi->num_locks) ? CHECK_FILELOCK : 0;
atomic_read       345 fs/ceph/dir.c  		int shared_gen = atomic_read(&ci->i_shared_gen);
atomic_read       766 fs/ceph/dir.c  			di->lease_shared_gen = atomic_read(&ci->i_shared_gen);
atomic_read      1507 fs/ceph/dir.c  		if (atomic_read(&ci->i_shared_gen) == di->lease_shared_gen &&
atomic_read      1531 fs/ceph/dir.c  	shared_gen = atomic_read(&ci->i_shared_gen);
atomic_read      1545 fs/ceph/dir.c  	     dir, (unsigned)atomic_read(&ci->i_shared_gen), dentry, valid);
atomic_read      1725 fs/ceph/dir.c  	    di->lease_shared_gen == atomic_read(&dir_ci->i_shared_gen))
atomic_read      1060 fs/ceph/inode.c 	di->lease_shared_gen = atomic_read(&ceph_inode(dir)->i_shared_gen);
atomic_read      1671 fs/ceph/inode.c 			    atomic_read(&ci->i_shared_gen)) {
atomic_read      1450 fs/ceph/mds_client.c 		if (atomic_read(&ci->i_filelock_ref) > 0) {
atomic_read      1752 fs/ceph/mds_client.c 		if (atomic_read(&ci->i_filelock_ref) > 0)
atomic_read      1778 fs/ceph/mds_client.c 			count = atomic_read(&inode->i_count);
atomic_read      4334 fs/ceph/mds_client.c 	return atomic_read(&mdsc->num_sessions) <= skipped;
atomic_read        69 fs/ceph/snap.c 	     atomic_read(&realm->nref), atomic_read(&realm->nref)+1);
atomic_read       202 fs/ceph/snap.c 	     atomic_read(&realm->nref), atomic_read(&realm->nref)-1);
atomic_read       214 fs/ceph/snap.c 	     atomic_read(&realm->nref), atomic_read(&realm->nref)-1);
atomic_read      1154 fs/ceph/snap.c 		if (WARN_ON_ONCE(atomic_read(&sm->ref))) {
atomic_read       302 fs/cifs/cifs_debug.c 			atomic_read(&server->smbd_conn->send_credits),
atomic_read       303 fs/cifs/cifs_debug.c 			atomic_read(&server->smbd_conn->receive_credits),
atomic_read       307 fs/cifs/cifs_debug.c 			atomic_read(&server->smbd_conn->send_pending),
atomic_read       308 fs/cifs/cifs_debug.c 			atomic_read(&server->smbd_conn->send_payload_pending));
atomic_read       319 fs/cifs/cifs_debug.c 			atomic_read(&server->smbd_conn->mr_ready_count),
atomic_read       320 fs/cifs/cifs_debug.c 			atomic_read(&server->smbd_conn->mr_used_count));
atomic_read       370 fs/cifs/cifs_debug.c 				atomic_read(&server->in_send),
atomic_read       371 fs/cifs/cifs_debug.c 				atomic_read(&server->num_waiters));
atomic_read       517 fs/cifs/cifs_debug.c 				atomic_read(&totBufAllocCount),
atomic_read       518 fs/cifs/cifs_debug.c 				atomic_read(&totSmBufAllocCount));
atomic_read       521 fs/cifs/cifs_debug.c 	seq_printf(m, "Operations (MIDs): %d\n", atomic_read(&midCount));
atomic_read       543 fs/cifs/cifs_debug.c 				atomic_read(&server->num_cmds[j]),
atomic_read       548 fs/cifs/cifs_debug.c 			if (atomic_read(&server->smb2slowcmd[j]))
atomic_read       550 fs/cifs/cifs_debug.c 					atomic_read(&server->smb2slowcmd[j]),
atomic_read       565 fs/cifs/cifs_debug.c 					   atomic_read(&tcon->num_smbs_sent));
atomic_read      1279 fs/cifs/connect.c 					 atomic_read(&midCount));
atomic_read      5492 fs/cifs/connect.c 		    atomic_read(&tlink->tl_count) != 0 ||
atomic_read       645 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_oplock_brks));
atomic_read       647 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_reads),
atomic_read       650 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_writes),
atomic_read       653 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_flushes));
atomic_read       655 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_locks),
atomic_read       656 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_hardlinks),
atomic_read       657 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_symlinks));
atomic_read       659 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_opens),
atomic_read       660 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_closes),
atomic_read       661 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_deletes));
atomic_read       663 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_posixopens),
atomic_read       664 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_posixmkdirs));
atomic_read       666 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_mkdirs),
atomic_read       667 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_rmdirs));
atomic_read       669 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_renames),
atomic_read       670 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_t2renames));
atomic_read       672 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_ffirst),
atomic_read       673 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_fnext),
atomic_read       674 fs/cifs/smb1ops.c 		   atomic_read(&tcon->stats.cifs_stats.num_fclose));
atomic_read      1292 fs/cifs/smb2ops.c 		   atomic_read(&tcon->num_local_opens),
atomic_read      1293 fs/cifs/smb2ops.c 		   atomic_read(&tcon->num_remote_opens));
atomic_read      1295 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_TREE_CONNECT_HE]),
atomic_read      1296 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_TREE_CONNECT_HE]));
atomic_read      1298 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_TREE_DISCONNECT_HE]),
atomic_read      1299 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_TREE_DISCONNECT_HE]));
atomic_read      1301 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_CREATE_HE]),
atomic_read      1302 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_CREATE_HE]));
atomic_read      1304 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_CLOSE_HE]),
atomic_read      1305 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_CLOSE_HE]));
atomic_read      1307 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_FLUSH_HE]),
atomic_read      1308 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_FLUSH_HE]));
atomic_read      1310 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_READ_HE]),
atomic_read      1311 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_READ_HE]));
atomic_read      1313 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_WRITE_HE]),
atomic_read      1314 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_WRITE_HE]));
atomic_read      1316 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_LOCK_HE]),
atomic_read      1317 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_LOCK_HE]));
atomic_read      1319 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_IOCTL_HE]),
atomic_read      1320 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_IOCTL_HE]));
atomic_read      1322 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_QUERY_DIRECTORY_HE]),
atomic_read      1323 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_QUERY_DIRECTORY_HE]));
atomic_read      1325 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_CHANGE_NOTIFY_HE]),
atomic_read      1326 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_CHANGE_NOTIFY_HE]));
atomic_read      1328 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_QUERY_INFO_HE]),
atomic_read      1329 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_QUERY_INFO_HE]));
atomic_read      1331 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_SET_INFO_HE]),
atomic_read      1332 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_SET_INFO_HE]));
atomic_read      1334 fs/cifs/smb2ops.c 		   atomic_read(&sent[SMB2_OPLOCK_BREAK_HE]),
atomic_read      1335 fs/cifs/smb2ops.c 		   atomic_read(&failed[SMB2_OPLOCK_BREAK_HE]));
atomic_read       401 fs/cifs/smbdirect.c 	if (atomic_read(&info->receive_credits) <
atomic_read       423 fs/cifs/smbdirect.c 		atomic_read(&info->receive_credits)) {
atomic_read       471 fs/cifs/smbdirect.c 	if (atomic_read(&info->send_credits))
atomic_read       840 fs/cifs/smbdirect.c 		atomic_read(&info->send_credits) > 0 ||
atomic_read      1422 fs/cifs/smbdirect.c 		atomic_read(&info->send_pending) == 0);
atomic_read      1424 fs/cifs/smbdirect.c 		atomic_read(&info->send_payload_pending) == 0);
atomic_read      1457 fs/cifs/smbdirect.c 	while (atomic_read(&info->mr_used_count)) {
atomic_read      2239 fs/cifs/smbdirect.c 		atomic_read(&info->send_payload_pending) == 0);
atomic_read      2405 fs/cifs/smbdirect.c 		atomic_read(&info->mr_ready_count) ||
atomic_read       119 fs/cifs/transport.c 		if (atomic_read(&server->num_cmds[smb_cmd]) == 0) {
atomic_read        36 fs/coda/cache.c 	cii->c_cached_epoch = atomic_read(&permission_epoch);
atomic_read        50 fs/coda/cache.c 	cii->c_cached_epoch = atomic_read(&permission_epoch) - 1;
atomic_read        70 fs/coda/cache.c 	    cii->c_cached_epoch == atomic_read(&permission_epoch);
atomic_read       147 fs/configfs/configfs_internal.h 		WARN_ON(!atomic_read(&sd->s_count));
atomic_read       155 fs/configfs/configfs_internal.h 	WARN_ON(!atomic_read(&sd->s_count));
atomic_read       381 fs/coredump.c  	if (atomic_read(&mm->mm_users) == nr + 1)
atomic_read       780 fs/debugfs/file.c 	*val = atomic_read((atomic_t *)data);
atomic_read       920 fs/dlm/user.c  	return atomic_read(&dlm_monitor_opened) ? 1 : 0;
atomic_read       492 fs/ecryptfs/miscdev.c 	BUG_ON(atomic_read(&ecryptfs_num_miscdev_opens) != 0);
atomic_read       157 fs/erofs/internal.h 	if (val != atomic_read(&grp->refcount)) {
atomic_read       172 fs/erofs/internal.h 	int v = atomic_read(&grp->refcount);
atomic_read        94 fs/erofs/utils.c 	if (atomic_read(&grp->refcount) != 1) {
atomic_read      1326 fs/erofs/zdata.c 		   !atomic_read(&io[JQ_SUBMIT].pending_bios));
atomic_read       132 fs/erofs/zdata.h 	return atomic_read(u.o) >> Z_EROFS_ONLINEPAGE_INDEX_SHIFT;
atomic_read       185 fs/erofs/zdata.h 	erofs_dbg("%s, page %p value %x", __func__, page, atomic_read(u.o));
atomic_read      1738 fs/exec.c      	    atomic_read(&current_user()->processes) > rlimit(RLIMIT_NPROC)) {
atomic_read       225 fs/ext2/xattr.c 		atomic_read(&(bh->b_count)), le32_to_cpu(HDR(bh)->h_refcount));
atomic_read       311 fs/ext2/xattr.c 		atomic_read(&(bh->b_count)), le32_to_cpu(HDR(bh)->h_refcount));
atomic_read       448 fs/ext2/xattr.c 			atomic_read(&(bh->b_count)),
atomic_read       812 fs/ext2/xattr.c 	ea_bdebug(bh, "b_count=%d", atomic_read(&(bh->b_count)));
atomic_read       968 fs/ext2/xattr.c 					  atomic_read(&(bh->b_count)));
atomic_read      3072 fs/ext4/ext4.h 	return (atomic_read(&sbi->s_lock_busy) > EXT4_CONTENTION_THRESHOLD);
atomic_read      1771 fs/ext4/extents.c 	     atomic_read(&EXT4_I(inode)->i_unwritten) ||
atomic_read        94 fs/ext4/file.c 			(atomic_read(&inode->i_writecount) == 1) &&
atomic_read       111 fs/ext4/file.c 	wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_unwritten) == 0));
atomic_read       251 fs/ext4/ialloc.c 	if (atomic_read(&inode->i_count) > 1) {
atomic_read       254 fs/ext4/ialloc.c 			 atomic_read(&inode->i_count));
atomic_read       374 fs/ext4/ialloc.c 		stats->free_inodes = atomic_read(&fg->free_inodes);
atomic_read       376 fs/ext4/ialloc.c 		stats->used_dirs = atomic_read(&fg->used_dirs);
atomic_read      4243 fs/ext4/inode.c 				atomic_read(&page->_refcount) == 1,
atomic_read      2752 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_allocated),
atomic_read      2753 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_reqs),
atomic_read      2754 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_success));
atomic_read      2758 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_ex_scanned),
atomic_read      2759 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_goals),
atomic_read      2760 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_2orders),
atomic_read      2761 fs/ext4/mballoc.c 				atomic_read(&sbi->s_bal_breaks),
atomic_read      2762 fs/ext4/mballoc.c 				atomic_read(&sbi->s_mb_lost_chunks));
atomic_read      2769 fs/ext4/mballoc.c 				atomic_read(&sbi->s_mb_preallocated),
atomic_read      2770 fs/ext4/mballoc.c 				atomic_read(&sbi->s_mb_discarded));
atomic_read      3577 fs/ext4/mballoc.c 	BUG_ON(atomic_read(&pa->pa_count));
atomic_read      3941 fs/ext4/mballoc.c 		if (atomic_read(&pa->pa_count)) {
atomic_read      4039 fs/ext4/mballoc.c 		if (atomic_read(&pa->pa_count)) {
atomic_read      4306 fs/ext4/mballoc.c 		if (atomic_read(&pa->pa_count)) {
atomic_read      5198 fs/ext4/mballoc.c 	    minblocks >= atomic_read(&EXT4_SB(sb)->s_last_trim_minblks))
atomic_read       325 fs/ext4/sysfs.c 				atomic_read((atomic_t *) ptr));
atomic_read       533 fs/ext4/xattr.c 		atomic_read(&(bh->b_count)), le32_to_cpu(BHDR(bh)->h_refcount));
atomic_read       706 fs/ext4/xattr.c 		atomic_read(&(bh->b_count)), le32_to_cpu(BHDR(bh)->h_refcount));
atomic_read      1826 fs/ext4/xattr.c 			atomic_read(&(bs->bh->b_count)),
atomic_read      2310 fs/f2fs/data.c 			if (atomic_read(&sbi->wb_sync_req[DATA]) &&
atomic_read      2455 fs/f2fs/data.c 	else if (atomic_read(&sbi->wb_sync_req[DATA]))
atomic_read        45 fs/f2fs/debug.c 	si->ext_tree = atomic_read(&sbi->total_ext_tree);
atomic_read        46 fs/f2fs/debug.c 	si->zombie_tree = atomic_read(&sbi->total_zombie_tree);
atomic_read        47 fs/f2fs/debug.c 	si->ext_node = atomic_read(&sbi->total_ext_node);
atomic_read        59 fs/f2fs/debug.c 	si->aw_cnt = atomic_read(&sbi->aw_cnt);
atomic_read        60 fs/f2fs/debug.c 	si->vw_cnt = atomic_read(&sbi->vw_cnt);
atomic_read        61 fs/f2fs/debug.c 	si->max_aw_cnt = atomic_read(&sbi->max_aw_cnt);
atomic_read        62 fs/f2fs/debug.c 	si->max_vw_cnt = atomic_read(&sbi->max_vw_cnt);
atomic_read        72 fs/f2fs/debug.c 			atomic_read(&SM_I(sbi)->fcc_info->issued_flush);
atomic_read        74 fs/f2fs/debug.c 			atomic_read(&SM_I(sbi)->fcc_info->queued_flush);
atomic_read        80 fs/f2fs/debug.c 			atomic_read(&SM_I(sbi)->dcc_info->issued_discard);
atomic_read        82 fs/f2fs/debug.c 			atomic_read(&SM_I(sbi)->dcc_info->queued_discard);
atomic_read        84 fs/f2fs/debug.c 			atomic_read(&SM_I(sbi)->dcc_info->discard_cmd_cnt);
atomic_read        94 fs/f2fs/debug.c 	si->inline_xattr = atomic_read(&sbi->inline_xattr);
atomic_read        95 fs/f2fs/debug.c 	si->inline_inode = atomic_read(&sbi->inline_inode);
atomic_read        96 fs/f2fs/debug.c 	si->inline_dir = atomic_read(&sbi->inline_dir);
atomic_read       138 fs/f2fs/debug.c 		si->meta_count[i] = atomic_read(&sbi->meta_count[i]);
atomic_read       145 fs/f2fs/debug.c 	si->inplace_count = atomic_read(&sbi->inplace_count);
atomic_read       250 fs/f2fs/debug.c 			atomic_read(&SM_I(sbi)->dcc_info->discard_cmd_cnt);
atomic_read       263 fs/f2fs/debug.c 	si->cache_mem += atomic_read(&sbi->total_ext_tree) *
atomic_read       265 fs/f2fs/debug.c 	si->cache_mem += atomic_read(&sbi->total_ext_node) *
atomic_read       304 fs/f2fs/extent_cache.c 	unsigned int count = atomic_read(&et->node_cnt);
atomic_read       314 fs/f2fs/extent_cache.c 	return count - atomic_read(&et->node_cnt);
atomic_read       352 fs/f2fs/extent_cache.c 	if (atomic_read(&et->node_cnt))
atomic_read       641 fs/f2fs/extent_cache.c 	if (!atomic_read(&sbi->total_zombie_tree))
atomic_read       649 fs/f2fs/extent_cache.c 		if (atomic_read(&et->node_cnt)) {
atomic_read       654 fs/f2fs/extent_cache.c 		f2fs_bug_on(sbi, atomic_read(&et->node_cnt));
atomic_read       713 fs/f2fs/extent_cache.c 	if (!et || !atomic_read(&et->node_cnt))
atomic_read       755 fs/f2fs/extent_cache.c 					atomic_read(&et->node_cnt)) {
atomic_read       768 fs/f2fs/extent_cache.c 	f2fs_bug_on(sbi, atomic_read(&et->node_cnt));
atomic_read      1385 fs/f2fs/f2fs.h 	if (atomic_read(&ffi->inject_ops) >= ffi->inject_rate) {
atomic_read      1909 fs/f2fs/f2fs.h 	return atomic_read(&sbi->nr_pages[count_type]);
atomic_read      1914 fs/f2fs/f2fs.h 	return atomic_read(&F2FS_I(inode)->dirty_pages);
atomic_read      2248 fs/f2fs/f2fs.h 			atomic_read(&SM_I(sbi)->dcc_info->queued_discard))
atomic_read      2252 fs/f2fs/f2fs.h 			atomic_read(&SM_I(sbi)->fcc_info->queued_flush))
atomic_read      3381 fs/f2fs/f2fs.h 		int cur = atomic_read(&F2FS_I_SB(inode)->aw_cnt);	\
atomic_read      3382 fs/f2fs/f2fs.h 		int max = atomic_read(&F2FS_I_SB(inode)->max_aw_cnt);	\
atomic_read      3392 fs/f2fs/f2fs.h 		int cur = atomic_read(&F2FS_I_SB(inode)->vw_cnt);	\
atomic_read      3393 fs/f2fs/f2fs.h 		int max = atomic_read(&F2FS_I_SB(inode)->max_vw_cnt);	\
atomic_read      1635 fs/f2fs/file.c 			atomic_read(&inode->i_writecount) != 1)
atomic_read        84 fs/f2fs/node.c 		mem_size = (atomic_read(&sbi->total_ext_tree) *
atomic_read        86 fs/f2fs/node.c 				atomic_read(&sbi->total_ext_node) *
atomic_read      1834 fs/f2fs/node.c 			if (atomic_read(&sbi->wb_sync_req[NODE]) &&
atomic_read      1993 fs/f2fs/node.c 	else if (atomic_read(&sbi->wb_sync_req[NODE]))
atomic_read      1687 fs/f2fs/segment.c 	f2fs_bug_on(sbi, atomic_read(&dcc->discard_cmd_cnt));
atomic_read      1715 fs/f2fs/segment.c 		if (atomic_read(&dcc->queued_discard))
atomic_read      2091 fs/f2fs/segment.c 	if (unlikely(atomic_read(&dcc->discard_cmd_cnt)))
atomic_read        35 fs/f2fs/shrinker.c 	return atomic_read(&sbi->total_zombie_tree) +
atomic_read        36 fs/f2fs/shrinker.c 				atomic_read(&sbi->total_ext_node);
atomic_read       160 fs/file.c      	if (atomic_read(&files->count) > 1)
atomic_read       770 fs/file.c      	if (atomic_read(&files->count) == 1) {
atomic_read       208 fs/fs-writeback.c 	wait_event(*done->waitq, !atomic_read(&done->cnt));
atomic_read       500 fs/fs-writeback.c 	if (atomic_read(&isw_nr_in_flight) > WB_FRN_MAX_IN_FLIGHT)
atomic_read      1002 fs/fs-writeback.c 	if (atomic_read(&isw_nr_in_flight)) {
atomic_read      1534 fs/fs-writeback.c 	if (!atomic_read(&inode->i_count))
atomic_read       402 fs/fscache/cache.c 		   atomic_read(&cache->object_count) == 0);
atomic_read        38 fs/fscache/cookie.c 	       atomic_read(&cookie->n_children),
atomic_read        39 fs/fscache/cookie.c 	       atomic_read(&cookie->n_active));
atomic_read       221 fs/fscache/cookie.c 				     atomic_read(&cursor->usage));
atomic_read       720 fs/fscache/cookie.c 	ASSERTCMP(atomic_read(&cookie->n_active), >, 0);
atomic_read       722 fs/fscache/cookie.c 	if (atomic_read(&cookie->n_children) != 0) {
atomic_read       766 fs/fscache/cookie.c 			       !atomic_read(&cookie->n_active));
atomic_read       808 fs/fscache/cookie.c 	       atomic_read(&cookie->n_active), retire);
atomic_read       824 fs/fscache/cookie.c 		ASSERTCMP(atomic_read(&cookie->parent->usage), >, 0);
atomic_read       825 fs/fscache/cookie.c 		ASSERTCMP(atomic_read(&cookie->parent->n_children), >, 0);
atomic_read       830 fs/fscache/cookie.c 	ASSERTCMP(atomic_read(&cookie->usage), >, 0);
atomic_read        37 fs/fscache/histogram.c 		n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]);
atomic_read        38 fs/fscache/histogram.c 		n[1] = atomic_read(&fscache_ops_histogram[index]);
atomic_read        39 fs/fscache/histogram.c 		n[2] = atomic_read(&fscache_objs_histogram[index]);
atomic_read        40 fs/fscache/histogram.c 		n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]);
atomic_read        41 fs/fscache/histogram.c 		n[4] = atomic_read(&fscache_retrieval_histogram[index]);
atomic_read       225 fs/fscache/object-list.c 		FILTER(atomic_read(&obj->n_reads),
atomic_read       242 fs/fscache/object-list.c 		   atomic_read(&obj->n_reads),
atomic_read        63 fs/fscache/operation.c 	       op->object->debug_id, op->debug_id, atomic_read(&op->usage));
atomic_read        68 fs/fscache/operation.c 	ASSERTCMP(atomic_read(&op->usage), >, 0);
atomic_read       166 fs/fscache/operation.c 	ASSERTCMP(atomic_read(&op->usage), >, 0);
atomic_read       248 fs/fscache/operation.c 	       object->debug_id, op->debug_id, atomic_read(&op->usage));
atomic_read       253 fs/fscache/operation.c 	ASSERTCMP(atomic_read(&op->usage), >, 0);
atomic_read       373 fs/fscache/operation.c 	ASSERTCMP(atomic_read(&op->usage), >, 0);
atomic_read       500 fs/fscache/operation.c 	       op->debug_id, atomic_read(&op->usage));
atomic_read       502 fs/fscache/operation.c 	ASSERTCMP(atomic_read(&op->usage), >, 0);
atomic_read       592 fs/fscache/operation.c 		ASSERTCMP(atomic_read(&op->usage), ==, 0);
atomic_read       622 fs/fscache/operation.c 	       op->object->debug_id, op->debug_id, atomic_read(&op->usage));
atomic_read       290 fs/fscache/page.c 		    atomic_read(&op->n_pages), ==, 0);
atomic_read       799 fs/fscache/page.c 	_enter("{OP%x,%d}", op->op.debug_id, atomic_read(&op->op.usage));
atomic_read       142 fs/fscache/stats.c 		   atomic_read(&fscache_n_cookie_index),
atomic_read       143 fs/fscache/stats.c 		   atomic_read(&fscache_n_cookie_data),
atomic_read       144 fs/fscache/stats.c 		   atomic_read(&fscache_n_cookie_special));
atomic_read       147 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_alloc),
atomic_read       148 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_no_alloc),
atomic_read       149 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_avail),
atomic_read       150 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_dead));
atomic_read       152 fs/fscache/stats.c 		   atomic_read(&fscache_n_checkaux_none),
atomic_read       153 fs/fscache/stats.c 		   atomic_read(&fscache_n_checkaux_okay),
atomic_read       154 fs/fscache/stats.c 		   atomic_read(&fscache_n_checkaux_update),
atomic_read       155 fs/fscache/stats.c 		   atomic_read(&fscache_n_checkaux_obsolete));
atomic_read       158 fs/fscache/stats.c 		   atomic_read(&fscache_n_marks),
atomic_read       159 fs/fscache/stats.c 		   atomic_read(&fscache_n_uncaches));
atomic_read       163 fs/fscache/stats.c 		   atomic_read(&fscache_n_acquires),
atomic_read       164 fs/fscache/stats.c 		   atomic_read(&fscache_n_acquires_null),
atomic_read       165 fs/fscache/stats.c 		   atomic_read(&fscache_n_acquires_no_cache),
atomic_read       166 fs/fscache/stats.c 		   atomic_read(&fscache_n_acquires_ok),
atomic_read       167 fs/fscache/stats.c 		   atomic_read(&fscache_n_acquires_nobufs),
atomic_read       168 fs/fscache/stats.c 		   atomic_read(&fscache_n_acquires_oom));
atomic_read       171 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_lookups),
atomic_read       172 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_lookups_negative),
atomic_read       173 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_lookups_positive),
atomic_read       174 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_created),
atomic_read       175 fs/fscache/stats.c 		   atomic_read(&fscache_n_object_lookups_timed_out));
atomic_read       178 fs/fscache/stats.c 		   atomic_read(&fscache_n_invalidates),
atomic_read       179 fs/fscache/stats.c 		   atomic_read(&fscache_n_invalidates_run));
atomic_read       182 fs/fscache/stats.c 		   atomic_read(&fscache_n_updates),
atomic_read       183 fs/fscache/stats.c 		   atomic_read(&fscache_n_updates_null),
atomic_read       184 fs/fscache/stats.c 		   atomic_read(&fscache_n_updates_run));
atomic_read       187 fs/fscache/stats.c 		   atomic_read(&fscache_n_relinquishes),
atomic_read       188 fs/fscache/stats.c 		   atomic_read(&fscache_n_relinquishes_null),
atomic_read       189 fs/fscache/stats.c 		   atomic_read(&fscache_n_relinquishes_waitcrt),
atomic_read       190 fs/fscache/stats.c 		   atomic_read(&fscache_n_relinquishes_retire));
atomic_read       193 fs/fscache/stats.c 		   atomic_read(&fscache_n_attr_changed),
atomic_read       194 fs/fscache/stats.c 		   atomic_read(&fscache_n_attr_changed_ok),
atomic_read       195 fs/fscache/stats.c 		   atomic_read(&fscache_n_attr_changed_nobufs),
atomic_read       196 fs/fscache/stats.c 		   atomic_read(&fscache_n_attr_changed_nomem),
atomic_read       197 fs/fscache/stats.c 		   atomic_read(&fscache_n_attr_changed_calls));
atomic_read       200 fs/fscache/stats.c 		   atomic_read(&fscache_n_allocs),
atomic_read       201 fs/fscache/stats.c 		   atomic_read(&fscache_n_allocs_ok),
atomic_read       202 fs/fscache/stats.c 		   atomic_read(&fscache_n_allocs_wait),
atomic_read       203 fs/fscache/stats.c 		   atomic_read(&fscache_n_allocs_nobufs),
atomic_read       204 fs/fscache/stats.c 		   atomic_read(&fscache_n_allocs_intr));
atomic_read       206 fs/fscache/stats.c 		   atomic_read(&fscache_n_alloc_ops),
atomic_read       207 fs/fscache/stats.c 		   atomic_read(&fscache_n_alloc_op_waits),
atomic_read       208 fs/fscache/stats.c 		   atomic_read(&fscache_n_allocs_object_dead));
atomic_read       212 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals),
atomic_read       213 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_ok),
atomic_read       214 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_wait),
atomic_read       215 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_nodata),
atomic_read       216 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_nobufs),
atomic_read       217 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_intr),
atomic_read       218 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_nomem));
atomic_read       220 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrieval_ops),
atomic_read       221 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrieval_op_waits),
atomic_read       222 fs/fscache/stats.c 		   atomic_read(&fscache_n_retrievals_object_dead));
atomic_read       225 fs/fscache/stats.c 		   atomic_read(&fscache_n_stores),
atomic_read       226 fs/fscache/stats.c 		   atomic_read(&fscache_n_stores_ok),
atomic_read       227 fs/fscache/stats.c 		   atomic_read(&fscache_n_stores_again),
atomic_read       228 fs/fscache/stats.c 		   atomic_read(&fscache_n_stores_nobufs),
atomic_read       229 fs/fscache/stats.c 		   atomic_read(&fscache_n_stores_oom));
atomic_read       231 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_ops),
atomic_read       232 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_calls),
atomic_read       233 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_pages),
atomic_read       234 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_radix_deletes),
atomic_read       235 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_pages_over_limit));
atomic_read       238 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_vmscan_not_storing),
atomic_read       239 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_vmscan_gone),
atomic_read       240 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_vmscan_busy),
atomic_read       241 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_vmscan_cancelled),
atomic_read       242 fs/fscache/stats.c 		   atomic_read(&fscache_n_store_vmscan_wait));
atomic_read       245 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_pend),
atomic_read       246 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_run),
atomic_read       247 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_enqueue),
atomic_read       248 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_cancelled),
atomic_read       249 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_rejected));
atomic_read       251 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_initialised),
atomic_read       252 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_deferred_release),
atomic_read       253 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_release),
atomic_read       254 fs/fscache/stats.c 		   atomic_read(&fscache_n_op_gc));
atomic_read       257 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_alloc_object),
atomic_read       258 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_lookup_object),
atomic_read       259 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_lookup_complete),
atomic_read       260 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_grab_object));
atomic_read       262 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_invalidate_object),
atomic_read       263 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_update_object),
atomic_read       264 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_drop_object),
atomic_read       265 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_put_object),
atomic_read       266 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_attr_changed),
atomic_read       267 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_sync_cache));
atomic_read       269 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_read_or_alloc_page),
atomic_read       270 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_read_or_alloc_pages),
atomic_read       271 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_allocate_page),
atomic_read       272 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_allocate_pages),
atomic_read       273 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_write_page),
atomic_read       274 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_uncache_page),
atomic_read       275 fs/fscache/stats.c 		   atomic_read(&fscache_n_cop_dissociate_pages));
atomic_read       277 fs/fscache/stats.c 		   atomic_read(&fscache_n_cache_no_space_reject),
atomic_read       278 fs/fscache/stats.c 		   atomic_read(&fscache_n_cache_stale_objects),
atomic_read       279 fs/fscache/stats.c 		   atomic_read(&fscache_n_cache_retired_objects),
atomic_read       280 fs/fscache/stats.c 		   atomic_read(&fscache_n_cache_culled_objects));
atomic_read        59 fs/fuse/control.c 		value = atomic_read(&fc->num_waiting);
atomic_read       583 fs/fuse/cuse.c 	return sprintf(buf, "%d\n", atomic_read(&cc->fc.num_waiting));
atomic_read      2157 fs/fuse/dev.c  	wait_event(fc->blocked_waitq, atomic_read(&fc->num_waiting) == 0);
atomic_read       794 fs/gfs2/aops.c 		if (atomic_read(&bh->b_count))
atomic_read      1557 fs/gfs2/bmap.c 			if (isize_blks > atomic_read(&sdp->sd_log_thresh2))
atomic_read      1559 fs/gfs2/bmap.c 					atomic_read(&sdp->sd_log_thresh2);
atomic_read      1575 fs/gfs2/bmap.c 		    RES_QUOTA >= atomic_read(&sdp->sd_log_thresh2)) {
atomic_read       403 fs/gfs2/file.c 	if (hint > atomic_read(&ip->i_sizehint))
atomic_read       140 fs/gfs2/glock.c 	BUG_ON(atomic_read(&gl->gl_revokes));
atomic_read      1625 fs/gfs2/glock.c 	return vfs_pressure_ratio(atomic_read(&lru_count));
atomic_read      1737 fs/gfs2/glock.c 			   atomic_read(&sdp->sd_glock_disposal) == 0,
atomic_read      1905 fs/gfs2/glock.c 		  atomic_read(&gl->gl_ail_count),
atomic_read      1906 fs/gfs2/glock.c 		  atomic_read(&gl->gl_revokes),
atomic_read        76 fs/gfs2/glops.c 	GLOCK_BUG_ON(gl, !fsync && atomic_read(&gl->gl_ail_count));
atomic_read        90 fs/gfs2/glops.c 	tr.tr_revokes = atomic_read(&gl->gl_ail_count);
atomic_read       110 fs/gfs2/glops.c 		log_in_flight = atomic_read(&sdp->sd_log_in_flight);
atomic_read       140 fs/gfs2/glops.c 	unsigned int revokes = atomic_read(&gl->gl_ail_count);
atomic_read       219 fs/gfs2/glops.c 	gfs2_assert_withdraw(sdp, !atomic_read(&gl->gl_ail_count));
atomic_read       319 fs/gfs2/glops.c 	gfs2_assert_withdraw(gl->gl_name.ln_sbd, !atomic_read(&gl->gl_ail_count));
atomic_read       328 fs/gfs2/log.c  	gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <=
atomic_read       367 fs/gfs2/log.c  	free_blocks = atomic_read(&sdp->sd_log_blks_free);
atomic_read       374 fs/gfs2/log.c  			if (atomic_read(&sdp->sd_log_blks_free) <= wanted)
atomic_read       376 fs/gfs2/log.c  			free_blocks = atomic_read(&sdp->sd_log_blks_free);
atomic_read       509 fs/gfs2/log.c  	gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <=
atomic_read       520 fs/gfs2/log.c  	if (atomic_read(&sdp->sd_log_in_flight)) {
atomic_read       524 fs/gfs2/log.c  			if (atomic_read(&sdp->sd_log_in_flight))
atomic_read       526 fs/gfs2/log.c  		} while(atomic_read(&sdp->sd_log_in_flight));
atomic_read       766 fs/gfs2/log.c  	enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state);
atomic_read       795 fs/gfs2/log.c  	enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state);
atomic_read       921 fs/gfs2/log.c  	gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <=
atomic_read       947 fs/gfs2/log.c  	if (atomic_read(&sdp->sd_log_pinned) > atomic_read(&sdp->sd_log_thresh1) ||
atomic_read       948 fs/gfs2/log.c  	    ((sdp->sd_jdesc->jd_blocks - atomic_read(&sdp->sd_log_blks_free)) >
atomic_read       949 fs/gfs2/log.c  	    atomic_read(&sdp->sd_log_thresh2)))
atomic_read       978 fs/gfs2/log.c  	return (atomic_read(&sdp->sd_log_pinned) +
atomic_read       979 fs/gfs2/log.c  		atomic_read(&sdp->sd_log_blks_needed) >=
atomic_read       980 fs/gfs2/log.c  		atomic_read(&sdp->sd_log_thresh1));
atomic_read       985 fs/gfs2/log.c  	unsigned int used_blocks = sdp->sd_jdesc->jd_blocks - atomic_read(&sdp->sd_log_blks_free);
atomic_read       990 fs/gfs2/log.c  	return used_blocks + atomic_read(&sdp->sd_log_blks_needed) >=
atomic_read       991 fs/gfs2/log.c  		atomic_read(&sdp->sd_log_thresh2);
atomic_read       725 fs/gfs2/ops_fstype.c 	trace_gfs2_log_blocks(sdp, atomic_read(&sdp->sd_log_blks_free));
atomic_read       553 fs/gfs2/quota.c 	if (ip->i_qadata && ((wcount == NULL) || (atomic_read(wcount) <= 1))) {
atomic_read      1467 fs/gfs2/quota.c 	gfs2_assert_warn(sdp, !atomic_read(&sdp->sd_quota_count));
atomic_read       693 fs/gfs2/rgrp.c 	if ((wcount == NULL) || (atomic_read(wcount) <= 1))
atomic_read      1577 fs/gfs2/rgrp.c 		extlen = max_t(u32, atomic_read(&ip->i_sizehint), ap->target);
atomic_read       621 fs/gfs2/super.c 	wait_event(sdp->sd_reserving_log_wait, atomic_read(&sdp->sd_reserving_log) == 0);
atomic_read       622 fs/gfs2/super.c 	gfs2_assert_warn(sdp, atomic_read(&sdp->sd_log_blks_free) == sdp->sd_jdesc->jd_blocks);
atomic_read       761 fs/gfs2/super.c 	if (atomic_read(&sdp->sd_freeze_state) != SFS_UNFROZEN)
atomic_read       799 fs/gfs2/super.c         if (atomic_read(&sdp->sd_freeze_state) != SFS_FROZEN ||
atomic_read      1148 fs/gfs2/super.c 	if (atomic_read(&gl->gl_revokes) == 0) {
atomic_read       194 fs/gfs2/trans.c 	enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state);
atomic_read       302 fs/hfs/bnode.c 		node->tree->cnid, node->this, atomic_read(&node->refcnt));
atomic_read       449 fs/hfs/bnode.c 			atomic_read(&node->refcnt));
atomic_read       462 fs/hfs/bnode.c 			atomic_read(&node->refcnt));
atomic_read       463 fs/hfs/bnode.c 		BUG_ON(!atomic_read(&node->refcnt));
atomic_read       148 fs/hfs/btree.c 			if (atomic_read(&node->refcnt))
atomic_read       151 fs/hfs/btree.c 				       atomic_read(&node->refcnt));
atomic_read       102 fs/hfs/inode.c 		else if (atomic_read(&node->refcnt))
atomic_read       117 fs/hfs/inode.c 			if (atomic_read(&node->refcnt)) {
atomic_read       468 fs/hfsplus/bnode.c 		node->tree->cnid, node->this, atomic_read(&node->refcnt));
atomic_read       617 fs/hfsplus/bnode.c 			atomic_read(&node->refcnt));
atomic_read       630 fs/hfsplus/bnode.c 			atomic_read(&node->refcnt));
atomic_read       631 fs/hfsplus/bnode.c 		BUG_ON(!atomic_read(&node->refcnt));
atomic_read       269 fs/hfsplus/btree.c 			if (atomic_read(&node->refcnt))
atomic_read       273 fs/hfsplus/btree.c 					atomic_read(&node->refcnt));
atomic_read       373 fs/hfsplus/dir.c 	    atomic_read(&HFSPLUS_I(inode)->opencnt)) {
atomic_read       396 fs/hfsplus/dir.c 			if (!atomic_read(&HFSPLUS_I(inode)->opencnt)) {
atomic_read        98 fs/hfsplus/inode.c 		else if (atomic_read(&node->refcnt))
atomic_read       114 fs/hfsplus/inode.c 			if (atomic_read(&node->refcnt)) {
atomic_read       138 fs/hfsplus/xattr.c 	switch (atomic_read(&sbi->attr_tree_state)) {
atomic_read       187 fs/hpfs/inode.c 	if (hpfs_inode->i_rddir_off && !atomic_read(&i->i_count)) {
atomic_read       442 fs/inode.c     	    !atomic_read(&inode->i_count) && inode->i_sb->s_flags & SB_ACTIVE)
atomic_read       632 fs/inode.c     		if (atomic_read(&inode->i_count))
atomic_read       693 fs/inode.c     		if (atomic_read(&inode->i_count)) {
atomic_read       749 fs/inode.c     	if (atomic_read(&inode->i_count) ||
atomic_read      2103 fs/inode.c     		if (atomic_read(&inode->i_dio_count))
atomic_read      2105 fs/inode.c     	} while (atomic_read(&inode->i_dio_count));
atomic_read      2121 fs/inode.c     	if (atomic_read(&inode->i_dio_count))
atomic_read       435 fs/io_uring.c  					+ atomic_read(&ctx->cached_cq_overflow);
atomic_read      2340 fs/io_uring.c  	if (!atomic_read(&list->cnt))
atomic_read      2350 fs/io_uring.c  	if (!atomic_read(&list->cnt)) {
atomic_read      2951 fs/io_uring.c  			atomic_read(&ctx->cq_timeouts) != iowq->nr_timeouts;
atomic_read      3002 fs/io_uring.c  	iowq.nr_timeouts = atomic_read(&ctx->cq_timeouts);
atomic_read        52 fs/iomap/buffered-io.c 	WARN_ON_ONCE(atomic_read(&iop->read_count));
atomic_read        53 fs/iomap/buffered-io.c 	WARN_ON_ONCE(atomic_read(&iop->write_count));
atomic_read       671 fs/jbd2/checkpoint.c 	J_ASSERT(atomic_read(&transaction->t_updates) == 0);
atomic_read        69 fs/jbd2/commit.c 	if (atomic_read(&bh->b_count) != 1)
atomic_read       438 fs/jbd2/commit.c 	while (atomic_read(&commit_transaction->t_updates)) {
atomic_read       443 fs/jbd2/commit.c 		if (atomic_read(&commit_transaction->t_updates)) {
atomic_read       456 fs/jbd2/commit.c 	J_ASSERT (atomic_read(&commit_transaction->t_outstanding_credits) <=
atomic_read       518 fs/jbd2/commit.c 	atomic_sub(atomic_read(&journal->j_reserved_credits),
atomic_read       564 fs/jbd2/commit.c 		atomic_read(&commit_transaction->t_outstanding_credits);
atomic_read       568 fs/jbd2/commit.c 		 atomic_read(&commit_transaction->t_outstanding_credits));
atomic_read       824 fs/jbd2/commit.c 		J_ASSERT_BH(bh, atomic_read(&bh->b_count) == 0);
atomic_read      1093 fs/jbd2/commit.c 		atomic_read(&commit_transaction->t_handle_count);
atomic_read      2477 fs/jbd2/journal.c 			(atomic_read(&bh->b_count) > 0) ||
atomic_read      2717 fs/jbd2/journal.c 	int n = atomic_read(&nr_journal_heads);
atomic_read        91 fs/jbd2/transaction.c 		   atomic_read(&journal->j_reserved_credits));
atomic_read       236 fs/jbd2/transaction.c 		if (atomic_read(&journal->j_reserved_credits) + total >
atomic_read       241 fs/jbd2/transaction.c 				   atomic_read(&journal->j_reserved_credits) + total <=
atomic_read       284 fs/jbd2/transaction.c 			 atomic_read(&journal->j_reserved_credits) + rsv_blocks
atomic_read       413 fs/jbd2/transaction.c 		  atomic_read(&transaction->t_outstanding_credits),
atomic_read       692 fs/jbd2/transaction.c 	J_ASSERT(atomic_read(&transaction->t_updates) > 0);
atomic_read       756 fs/jbd2/transaction.c 	if (atomic_read(&journal->j_reserved_credits)) {
atomic_read       759 fs/jbd2/transaction.c 			   atomic_read(&journal->j_reserved_credits) == 0);
atomic_read       773 fs/jbd2/transaction.c 		if (!atomic_read(&transaction->t_updates)) {
atomic_read      1736 fs/jbd2/transaction.c 		J_ASSERT(atomic_read(&transaction->t_updates) > 0);
atomic_read      1822 fs/jbd2/transaction.c 	    (atomic_read(&transaction->t_outstanding_credits) >
atomic_read       865 fs/jffs2/xattr.c 			if (!atomic_read(&xd->refcnt)) {
atomic_read      1325 fs/jffs2/xattr.c 	if (atomic_read(&xd->refcnt) || xd->node != (void *)xd)
atomic_read       609 fs/jfs/jfs_dmap.c 	if ((atomic_read(&bmp->db_active[agpref]) == 0) &&
atomic_read       620 fs/jfs/jfs_dmap.c 		if (atomic_read(&bmp->db_active[agpref]))
atomic_read       755 fs/jfs/jfs_dmap.c 		if (atomic_read(&bmp->db_active[agno]))
atomic_read       789 fs/jfs/jfs_dmap.c 		writers = atomic_read(&bmp->db_active[agno]);
atomic_read       229 fs/jfs/jfs_imap.c 	dinom_le->in_numinos = cpu_to_le32(atomic_read(&imp->im_numinos));
atomic_read       230 fs/jfs/jfs_imap.c 	dinom_le->in_numfree = cpu_to_le32(atomic_read(&imp->im_numfree));
atomic_read      1361 fs/jfs/jfs_imap.c 	if (atomic_read(&JFS_SBI(pip->i_sb)->bmap->db_active[agno])) {
atomic_read      2851 fs/jfs/jfs_imap.c 		   imap->im_nextiag, atomic_read(&imap->im_numinos),
atomic_read      2852 fs/jfs/jfs_imap.c 		   atomic_read(&imap->im_numfree));
atomic_read      2957 fs/jfs/jfs_imap.c 	if (xnuminos != atomic_read(&imap->im_numinos) ||
atomic_read      2958 fs/jfs/jfs_imap.c 	    xnumfree != atomic_read(&imap->im_numfree)) {
atomic_read       140 fs/jfs/super.c 	maxinodes = min((s64) atomic_read(&imap->im_numinos) +
atomic_read       144 fs/jfs/super.c 	buf->f_ffree = maxinodes - (atomic_read(&imap->im_numinos) -
atomic_read       145 fs/jfs/super.c 				    atomic_read(&imap->im_numfree));
atomic_read        30 fs/kernfs/dir.c 	return atomic_read(&kn->active) >= 0;
atomic_read       469 fs/kernfs/dir.c 		if (atomic_read(&kn->active) != KN_DEACTIVATED_BIAS)
atomic_read       475 fs/kernfs/dir.c 		   atomic_read(&kn->active) == KN_DEACTIVATED_BIAS);
atomic_read       494 fs/kernfs/dir.c 		WARN_ON(!atomic_read(&kn->count));
atomic_read       525 fs/kernfs/dir.c 	WARN_ONCE(atomic_read(&kn->active) != KN_DEACTIVATED_BIAS,
atomic_read       527 fs/kernfs/dir.c 		  parent ? parent->name : "", kn->name, atomic_read(&kn->active));
atomic_read      1269 fs/kernfs/dir.c 		WARN_ON_ONCE(atomic_read(&pos->active) != KN_DEACTIVATED_BIAS);
atomic_read      1321 fs/kernfs/dir.c 			WARN_ON_ONCE(atomic_read(&kn->active) != KN_DEACTIVATED_BIAS);
atomic_read      1466 fs/kernfs/dir.c 			    atomic_read(&kn->active) == KN_DEACTIVATED_BIAS)
atomic_read       165 fs/kernfs/file.c 	of->event = atomic_read(&of->kn->attr.open->event);
atomic_read       210 fs/kernfs/file.c 	of->event = atomic_read(&of->kn->attr.open->event);
atomic_read       841 fs/kernfs/file.c 	if (of->event != atomic_read(&on->event))
atomic_read       366 fs/lockd/svc.c 	wait_event(nlm_ntf_wq, atomic_read(&nlm_ntf_refcnt) == 0);
atomic_read      1824 fs/locks.c     	if (atomic_read(&inode->i_writecount) != self_wcount ||
atomic_read      1825 fs/locks.c     	    atomic_read(&inode->i_readcount) != self_rcount)
atomic_read       409 fs/mbcache.c   		WARN_ON(atomic_read(&entry->e_refcnt) != 1);
atomic_read       455 fs/nfs/file.c  	if (atomic_read(&nfsi->commit_info.rpcs_out)) {
atomic_read       571 fs/nfs/inode.c 		atomic_read(&inode->i_count));
atomic_read      1809 fs/nfs/inode.c 			atomic_read(&inode->i_count), fattr->valid);
atomic_read       103 fs/nfs/pagelist.c 				       !atomic_read(&l_ctx->io_count));
atomic_read       121 fs/nfs/pagelist.c 	if (atomic_read(&l_ctx->io_count) > 0) {
atomic_read       126 fs/nfs/pagelist.c 	if (atomic_read(&l_ctx->io_count) == 0) {
atomic_read       527 fs/nfs/pnfs.c  		if (atomic_read(&lo->plh_outstanding) == 0)
atomic_read       929 fs/nfs/pnfs.c  		new_barrier = newseq - atomic_read(&lo->plh_outstanding);
atomic_read      1128 fs/nfs/pnfs.c  	if (atomic_read(&lo->plh_outstanding) != 0)
atomic_read      1950 fs/nfs/pnfs.c  	    atomic_read(&lo->plh_outstanding) != 0) {
atomic_read      1953 fs/nfs/pnfs.c  					!atomic_read(&lo->plh_outstanding)));
atomic_read        86 fs/nfs/pnfs_dev.c 			if (atomic_read(&d->ref))
atomic_read       332 fs/nfs/pnfs_dev.c 		if (d->nfs_client == clp && atomic_read(&d->ref)) {
atomic_read      1668 fs/nfs/write.c 				       !atomic_read(&cinfo->rpcs_out));
atomic_read      1993 fs/nfs/write.c 	if (!atomic_read(&nfsi->commit_info.rpcs_out))
atomic_read       378 fs/nfsd/filecache.c 	if (atomic_read(&nf->nf_ref) > 1)
atomic_read      1660 fs/nfsd/nfs4proc.c 	if (atomic_read(&ls->ls_stid.sc_file->fi_lo_recalls))
atomic_read       131 fs/nfsd/nfs4state.c 	if (atomic_read(&ses->se_ref) > ref_held_by_me)
atomic_read       582 fs/nfsd/nfs4state.c 		    atomic_read(&fp->fi_access[O_RDONLY]))
atomic_read       586 fs/nfsd/nfs4state.c 		    atomic_read(&fp->fi_access[O_WRONLY]))
atomic_read       601 fs/nfsd/nfs4state.c 		if (atomic_read(&fp->fi_access[1 - oflag]) == 0)
atomic_read      1946 fs/nfsd/nfs4state.c 		WARN_ON_ONCE(atomic_read(&ses->se_ref));
atomic_read      1996 fs/nfsd/nfs4state.c 	if (atomic_read(&clp->cl_rpc_users))
atomic_read      2552 fs/nfsd/nfs4state.c 	wait_event(expiry_wq, atomic_read(&clp->cl_rpc_users) == 0);
atomic_read      5194 fs/nfsd/nfs4state.c 			atomic_read(&nn->nr_reclaim_complete) ==
atomic_read       240 fs/nfsd/nfscache.c 		if (atomic_read(&nn->num_drc_entries) <= nn->max_drc_entries &&
atomic_read       277 fs/nfsd/nfscache.c 	return atomic_read(&nn->num_drc_entries);
atomic_read       368 fs/nfsd/nfscache.c 		nn->longest_chain_cachesize = atomic_read(&nn->num_drc_entries);
atomic_read       373 fs/nfsd/nfscache.c 				atomic_read(&nn->num_drc_entries));
atomic_read       578 fs/nfsd/nfscache.c 			atomic_read(&nn->num_drc_entries));
atomic_read       506 fs/nfsd/nfssvc.c 	wait_event(nn->ntf_wq, atomic_read(&nn->ntf_refcnt) == 0);
atomic_read       198 fs/nfsd/trace.h 		__entry->nf_ref = atomic_read(&nf->nf_ref);
atomic_read       247 fs/nfsd/trace.h 		__entry->nf_ref = nf ? atomic_read(&nf->nf_ref) : 0;
atomic_read       933 fs/nfsd/vfs.c  	if (atomic_read(&inode->i_writecount) > 1
atomic_read       180 fs/nilfs2/page.c 			       i++, bh, atomic_read(&bh->b_count),
atomic_read       350 fs/nilfs2/segbuf.c 		if (unlikely(atomic_read(&segbuf->sb_err))) {
atomic_read       507 fs/nilfs2/segbuf.c 	if (unlikely(atomic_read(&segbuf->sb_err) > 0)) {
atomic_read       280 fs/nilfs2/segment.c 		if (atomic_read(&nilfs->ns_ndirtyblks) > sci->sc_watermark)
atomic_read      1401 fs/nilfs2/segment.c 	if (atomic_read(&segbuf->sb_err)) {
atomic_read      1420 fs/nilfs2/segment.c 		if (atomic_read(&segbuf->sb_err) &&
atomic_read      2177 fs/nilfs2/segment.c 		if (atomic_read(&wait_req.done)) {
atomic_read      2199 fs/nilfs2/segment.c 		if (!atomic_read(&wrq->done) &&
atomic_read      2204 fs/nilfs2/segment.c 		if (atomic_read(&wrq->done)) {
atomic_read       632 fs/nilfs2/sysfs.c 	ndirtyblks = atomic_read(&nilfs->ns_ndirtyblks);
atomic_read       710 fs/nilfs2/the_nilfs.c 	nincsegs = atomic_read(&nilfs->ns_ndirtyblks) /
atomic_read       685 fs/notify/fanotify/fanotify_user.c 	if (atomic_read(&group->num_marks) > group->fanotify_data.max_marks)
atomic_read       803 fs/notify/fanotify/fanotify_user.c 	if (atomic_read(&user->fanotify_listeners) > FANOTIFY_DEFAULT_MAX_LISTENERS) {
atomic_read        64 fs/notify/fsnotify.c 		if (!atomic_read(&inode->i_count)) {
atomic_read        67 fs/notify/group.c 	wait_event(group->notification_waitq, !atomic_read(&group->user_waits));
atomic_read      2798 fs/ntfs/mft.c  	if (atomic_read(&ni->count) > 2) {
atomic_read      6095 fs/ocfs2/alloc.c 			atomic_read(&osb->osb_tl_disable) == 0) {
atomic_read       325 fs/ocfs2/cluster/heartbeat.c 	if (atomic_read(&reg->hr_steady_iterations) != 0)
atomic_read       855 fs/ocfs2/cluster/heartbeat.c 	if (atomic_read(&reg->hr_steady_iterations) != 0)
atomic_read      1169 fs/ocfs2/cluster/heartbeat.c 	if (atomic_read(&reg->hr_steady_iterations) != 0) {
atomic_read      1176 fs/ocfs2/cluster/heartbeat.c 	if (atomic_read(&reg->hr_steady_iterations) != 0) {
atomic_read      1880 fs/ocfs2/cluster/heartbeat.c 				atomic_read(&reg->hr_steady_iterations) == 0 ||
atomic_read      2101 fs/ocfs2/cluster/heartbeat.c 		       ((atomic_read(&reg->hr_steady_iterations) == 0) ?
atomic_read      2110 fs/ocfs2/cluster/heartbeat.c 	if (atomic_read(&reg->hr_steady_iterations) != 0) {
atomic_read       485 fs/ocfs2/cluster/tcp.c 	return atomic_read(&o2net_connected_peers);
atomic_read      1547 fs/ocfs2/cluster/tcp.c 	if (atomic_read(&nn->nn_timeout)) {
atomic_read      1603 fs/ocfs2/cluster/tcp.c 	timeout = atomic_read(&nn->nn_timeout);
atomic_read      1748 fs/ocfs2/cluster/tcp.c 	BUG_ON(atomic_read(&o2net_connected_peers) < 0);
atomic_read       102 fs/ocfs2/dlm/dlmdebug.c 	       res->inflight_locks, atomic_read(&res->asts_reserved));
atomic_read       508 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&res->asts_reserved),
atomic_read       722 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->res_cur_count),
atomic_read       723 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->res_tot_count));
atomic_read       726 fs/ocfs2/dlm/dlmdebug.c 		tot_mles += atomic_read(&dlm->mle_tot_count[i]);
atomic_read       729 fs/ocfs2/dlm/dlmdebug.c 		cur_mles += atomic_read(&dlm->mle_cur_count[i]);
atomic_read       738 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->mle_cur_count[DLM_MLE_BLOCK]),
atomic_read       739 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->mle_tot_count[DLM_MLE_BLOCK]));
atomic_read       744 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->mle_cur_count[DLM_MLE_MASTER]),
atomic_read       745 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->mle_tot_count[DLM_MLE_MASTER]));
atomic_read       750 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->mle_cur_count[DLM_MLE_MIGRATION]),
atomic_read       751 fs/ocfs2/dlm/dlmdebug.c 			atomic_read(&dlm->mle_tot_count[DLM_MLE_MIGRATION]));
atomic_read      1107 fs/ocfs2/dlm/dlmmaster.c 					 (atomic_read(&mle->woken) == 1),
atomic_read      2686 fs/ocfs2/dlm/dlmmaster.c 					(atomic_read(&mle->woken) == 1),
atomic_read      2690 fs/ocfs2/dlm/dlmmaster.c 		       	if (atomic_read(&mle->woken) == 1 ||
atomic_read       305 fs/ocfs2/journal.c 	flushed = atomic_read(&journal->j_num_trans);
atomic_read       323 fs/ocfs2/journal.c 	flushed = atomic_read(&journal->j_num_trans);
atomic_read       978 fs/ocfs2/journal.c 	num_running_trans = atomic_read(&(osb->journal->j_num_trans));
atomic_read       997 fs/ocfs2/journal.c 	BUG_ON(atomic_read(&(osb->journal->j_num_trans)) != 0);
atomic_read      1929 fs/ocfs2/journal.c 	if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE)
atomic_read      1933 fs/ocfs2/journal.c 					    atomic_read(&os->os_state));
atomic_read      1943 fs/ocfs2/journal.c 	if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE)
atomic_read      1965 fs/ocfs2/journal.c 					  atomic_read(&os->os_state));
atomic_read      1981 fs/ocfs2/journal.c 	if (atomic_read(&os->os_state) == ORPHAN_SCAN_ACTIVE)
atomic_read      1992 fs/ocfs2/journal.c 	if (atomic_read(&os->os_state) == ORPHAN_SCAN_ACTIVE) {
atomic_read      2274 fs/ocfs2/journal.c 		  (!quota && atomic_read(&osb->vol_state) == VOLUME_MOUNTED) ||
atomic_read      2275 fs/ocfs2/journal.c 		   atomic_read(&osb->vol_state) == VOLUME_MOUNTED_QUOTAS ||
atomic_read      2276 fs/ocfs2/journal.c 		   atomic_read(&osb->vol_state) == VOLUME_DISABLED);
atomic_read      2281 fs/ocfs2/journal.c 	if (atomic_read(&osb->vol_state) == VOLUME_DISABLED) {
atomic_read      2301 fs/ocfs2/journal.c 		 atomic_read(&journal->j_num_trans) == 0)) {
atomic_read      2304 fs/ocfs2/journal.c 					 atomic_read(&journal->j_num_trans)
atomic_read      2323 fs/ocfs2/journal.c 		if (kthread_should_stop() && atomic_read(&journal->j_num_trans)){
atomic_read      2327 fs/ocfs2/journal.c 			     atomic_read(&journal->j_num_trans));
atomic_read       215 fs/ocfs2/stack_user.c 	if ((c->oc_type == NO_CONTROLD) || atomic_read(&ocfs2_control_opened))
atomic_read      1037 fs/ocfs2/stack_user.c 		wait_event(lc->oc_wait, (atomic_read(&lc->oc_this_node) > 0));
atomic_read      1071 fs/ocfs2/stack_user.c 		rc = atomic_read(&lc->oc_this_node);
atomic_read       976 fs/ocfs2/suballoc.c 		atomic_read(&osb->s_num_meta_stolen) < OCFS2_MAX_TO_STEAL)
atomic_read      1065 fs/ocfs2/suballoc.c 	    atomic_read(&osb->s_num_inodes_stolen) < OCFS2_MAX_TO_STEAL)
atomic_read       230 fs/ocfs2/super.c 			atomic_read(&osb->vol_state), osb->osb_flags);
atomic_read       290 fs/ocfs2/super.c 			atomic_read(&osb->journal->j_num_trans));
atomic_read       296 fs/ocfs2/super.c 			atomic_read(&osb->alloc_stats.bitmap_data),
atomic_read       297 fs/ocfs2/super.c 			atomic_read(&osb->alloc_stats.local_data),
atomic_read       298 fs/ocfs2/super.c 			atomic_read(&osb->alloc_stats.bg_allocs),
atomic_read       299 fs/ocfs2/super.c 			atomic_read(&osb->alloc_stats.moves),
atomic_read       300 fs/ocfs2/super.c 			atomic_read(&osb->alloc_stats.bg_extends));
atomic_read       314 fs/ocfs2/super.c 			atomic_read(&osb->s_num_inodes_stolen),
atomic_read       316 fs/ocfs2/super.c 			atomic_read(&osb->s_num_meta_stolen));
atomic_read       323 fs/ocfs2/super.c 	if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE)
atomic_read        35 fs/overlayfs/inode.c 		if (atomic_read(&realinode->i_writecount) < 0)
atomic_read       287 fs/proc/array.c 		qsize = atomic_read(&__task_cred(p)->user->sigpending);
atomic_read      1079 fs/proc/base.c 			if (atomic_read(&p->mm->mm_users) > 1) {
atomic_read       228 fs/proc/generic.c 	if (atomic_read(&PDE(d_inode(dentry))->in_use) < 0)
atomic_read       235 fs/proc/generic.c 	return atomic_read(&PDE(d_inode(dentry))->in_use) < 0;
atomic_read       677 fs/proc/proc_sysctl.c 	if (event != atomic_read(&table->poll->event)) {
atomic_read        42 fs/proc/task_nommu.c 		if (atomic_read(&mm->mm_count) > 1 ||
atomic_read        52 fs/proc/task_nommu.c 	if (atomic_read(&mm->mm_count) > 1)
atomic_read        62 fs/proc/task_nommu.c 	if (current->files && atomic_read(&current->files->count) > 1)
atomic_read        44 fs/pstore/ram_core.c 	return atomic_read(&prz->buffer->size);
atomic_read        49 fs/pstore/ram_core.c 	return atomic_read(&prz->buffer->start);
atomic_read        62 fs/pstore/ram_core.c 	old = atomic_read(&prz->buffer->start);
atomic_read        84 fs/pstore/ram_core.c 	old = atomic_read(&prz->buffer->size);
atomic_read       548 fs/quota/dquot.c 		if (atomic_read(&dquot->dq_count)) {
atomic_read       560 fs/quota/dquot.c 				   atomic_read(&dquot->dq_count) == 1);
atomic_read       761 fs/quota/dquot.c 	if (!atomic_read(&dquot->dq_count)) {
atomic_read       771 fs/quota/dquot.c 	if (atomic_read(&dquot->dq_count) > 1) {
atomic_read       776 fs/quota/dquot.c 		    atomic_read(&dquot->dq_count) == 1)
atomic_read       888 fs/quota/dquot.c 		if (!atomic_read(&dquot->dq_count))
atomic_read       960 fs/quota/dquot.c 		    !atomic_read(&inode->i_writecount) ||
atomic_read      1445 fs/reiserfs/bitmap.c 		BUG_ON(atomic_read(&bh->b_count) == 0);
atomic_read      2318 fs/reiserfs/fix_node.c 		if (atomic_read(&(bh->b_count)) <= 0)
atomic_read       741 fs/reiserfs/journal.c 		if (atomic_read(&nr_reiserfs_jh) <= 0)
atomic_read       904 fs/reiserfs/journal.c 		    atomic_read(&other_jl->j_older_commits_done))
atomic_read       922 fs/reiserfs/journal.c 			if (atomic_read(&other_jl->j_commit_left) != 0) {
atomic_read       954 fs/reiserfs/journal.c 	if (atomic_read(&j->j_async_throttle)) {
atomic_read       986 fs/reiserfs/journal.c 	if (atomic_read(&jl->j_older_commits_done)) {
atomic_read      1018 fs/reiserfs/journal.c 	if (atomic_read(&jl->j_commit_left) <= 0) {
atomic_read      1099 fs/reiserfs/journal.c 	BUG_ON(atomic_read(&jl->j_commit_left) != 1);
atomic_read      1363 fs/reiserfs/journal.c 	if (atomic_read(&journal->j_wcount) != 0) {
atomic_read      1365 fs/reiserfs/journal.c 				 atomic_read(&journal->j_wcount));
atomic_read      1383 fs/reiserfs/journal.c 	if (atomic_read(&jl->j_nonzerolen) <= 0 &&
atomic_read      1384 fs/reiserfs/journal.c 	    atomic_read(&jl->j_commit_left) <= 0) {
atomic_read      1399 fs/reiserfs/journal.c 	if (atomic_read(&jl->j_nonzerolen) <= 0 &&
atomic_read      1400 fs/reiserfs/journal.c 	    atomic_read(&jl->j_commit_left) <= 0) {
atomic_read      1408 fs/reiserfs/journal.c 	if (atomic_read(&journal->j_wcount) != 0) {
atomic_read      1462 fs/reiserfs/journal.c 			if (atomic_read(&pjl->j_commit_left))
atomic_read      1516 fs/reiserfs/journal.c 			if (atomic_read(&saved_bh->b_count) < 0) {
atomic_read      1638 fs/reiserfs/journal.c 	if (jl->j_len == 0 || atomic_read(&jl->j_nonzerolen) == 0) {
atomic_read      1744 fs/reiserfs/journal.c 		    atomic_read(&jl->j_commit_left)
atomic_read      1805 fs/reiserfs/journal.c 		if (atomic_read(&tjl->j_commit_left) ||
atomic_read      1809 fs/reiserfs/journal.c 		cur_len = atomic_read(&tjl->j_nonzerolen);
atomic_read      2915 fs/reiserfs/journal.c 	    atomic_read(&journal->j_jlock) ||
atomic_read      2993 fs/reiserfs/journal.c 		while ((atomic_read(&journal->j_wcount) > 0 ||
atomic_read      2994 fs/reiserfs/journal.c 			atomic_read(&journal->j_jlock)) &&
atomic_read      3062 fs/reiserfs/journal.c 	    || (!join && atomic_read(&journal->j_wcount) > 0
atomic_read      3066 fs/reiserfs/journal.c 					      && atomic_read(&journal->j_jlock))
atomic_read      3077 fs/reiserfs/journal.c 			if (atomic_read(&journal->j_wcount) > 10) {
atomic_read      3087 fs/reiserfs/journal.c 		if (atomic_read(&journal->j_jlock)) {
atomic_read      3089 fs/reiserfs/journal.c 			       atomic_read(&journal->j_jlock)) {
atomic_read      3308 fs/reiserfs/journal.c 	if (atomic_read(&journal->j_wcount) <= 0) {
atomic_read      3311 fs/reiserfs/journal.c 				 atomic_read(&journal->j_wcount));
atomic_read      3449 fs/reiserfs/journal.c 		if (atomic_read(&bh->b_count) < 0) {
atomic_read      3498 fs/reiserfs/journal.c 		    atomic_read(&cur->jlist->j_commit_left) > 0 && cur->bh &&
atomic_read      3568 fs/reiserfs/journal.c 	if (atomic_read(&journal->j_wcount) <= 0 &&
atomic_read      3624 fs/reiserfs/journal.c 	if (atomic_read(&journal->j_wcount) > 0)
atomic_read      3643 fs/reiserfs/journal.c 	if (atomic_read(&journal->j_wcount) > 0) {
atomic_read      3662 fs/reiserfs/journal.c 				if (atomic_read(&journal->j_jlock)) {
atomic_read      3694 fs/reiserfs/journal.c 	if (!(journal->j_must_wait > 0) && !(atomic_read(&journal->j_jlock))
atomic_read      3792 fs/reiserfs/journal.c 						if (atomic_read
atomic_read      3883 fs/reiserfs/journal.c 			if (atomic_read(&jl->j_commit_left) > 1)
atomic_read       161 fs/reiserfs/prints.c 			 atomic_read(&(bh->b_count)),
atomic_read       673 fs/reiserfs/prints.c 			(tbSh) ? atomic_read(&tbSh->b_count) : -1,
atomic_read       675 fs/reiserfs/prints.c 			(tb->L[h]) ? atomic_read(&tb->L[h]->b_count) : -1,
atomic_read       677 fs/reiserfs/prints.c 			(tb->R[h]) ? atomic_read(&tb->R[h]->b_count) : -1,
atomic_read       719 fs/reiserfs/prints.c 			tb->FEB[i] ? atomic_read(&tb->FEB[i]->b_count) : 0,
atomic_read       117 fs/reiserfs/procfs.c 		   atomic_read(&r->s_generation_counter),
atomic_read       372 fs/reiserfs/procfs.c 		   atomic_read(&r->s_journal->j_wcount),
atomic_read      2305 fs/reiserfs/reiserfs.h #define get_generation(s) atomic_read (&fs_generation(s))
atomic_read       922 fs/reiserfs/stree.c #define held_by_others(bh) (atomic_read(&(bh)->b_count) > 1)
atomic_read      1554 fs/reiserfs/stree.c 	if (atomic_read(&inode->i_count) > 1 ||
atomic_read       346 fs/ubifs/super.c 	ubifs_assert(c, !atomic_read(&inode->i_count));
atomic_read       230 fs/udf/file.c  	    atomic_read(&inode->i_writecount) == 1) {
atomic_read        48 fs/xfs/libxfs/xfs_sb.c 		ASSERT(atomic_read(&pag->pag_ref) >= 0);
atomic_read        88 fs/xfs/libxfs/xfs_sb.c 	ASSERT(atomic_read(&pag->pag_ref) > 0);
atomic_read        72 fs/xfs/xfs_aops.c 	ASSERT(!iop || atomic_read(&iop->write_count) > 0);
atomic_read       881 fs/xfs/xfs_aops.c 	ASSERT(!iop || atomic_read(&iop->write_count) == 0);
atomic_read        52 fs/xfs/xfs_bmap_item.c 	ASSERT(atomic_read(&buip->bui_refcount) > 0);
atomic_read        87 fs/xfs/xfs_bmap_item.c 	ASSERT(atomic_read(&buip->bui_next_extent) ==
atomic_read       164 fs/xfs/xfs_buf.c 	ASSERT(atomic_read(&bp->b_hold) >= 1);
atomic_read       993 fs/xfs/xfs_buf.c 	ASSERT(atomic_read(&bp->b_hold) > 0);
atomic_read      1014 fs/xfs/xfs_buf.c 		if ((atomic_read(&bp->b_hold) == 1) && !list_empty(&bp->b_lru))
atomic_read      1021 fs/xfs/xfs_buf.c 	if (!(bp->b_flags & XBF_STALE) && atomic_read(&bp->b_lru_ref)) {
atomic_read      1101 fs/xfs/xfs_buf.c 	if (atomic_read(&bp->b_pin_count) && (bp->b_flags & XBF_STALE))
atomic_read      1124 fs/xfs/xfs_buf.c 	if (atomic_read(&bp->b_pin_count) == 0)
atomic_read      1130 fs/xfs/xfs_buf.c 		if (atomic_read(&bp->b_pin_count) == 0)
atomic_read      1567 fs/xfs/xfs_buf.c 	if (atomic_read(&bp->b_hold) > 1) {
atomic_read       311 fs/xfs/xfs_buf.h 	if (!list_empty(&bp->b_lru) || atomic_read(&bp->b_lru_ref) > 1)
atomic_read       318 fs/xfs/xfs_buf.h 	return atomic_read(&bp->b_pin_count);
atomic_read       127 fs/xfs/xfs_buf_item.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       306 fs/xfs/xfs_buf_item.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       365 fs/xfs/xfs_buf_item.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       401 fs/xfs/xfs_buf_item.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read        92 fs/xfs/xfs_dquot_item.c 	ASSERT(atomic_read(&dqp->q_pincount) > 0);
atomic_read       106 fs/xfs/xfs_dquot_item.c 	if (atomic_read(&dqp->q_pincount) == 0)
atomic_read       113 fs/xfs/xfs_dquot_item.c 	wait_event(dqp->q_pinwait, (atomic_read(&dqp->q_pincount) == 0));
atomic_read       145 fs/xfs/xfs_dquot_item.c 	if (atomic_read(&dqp->q_pincount) > 0)
atomic_read       170 fs/xfs/xfs_dquot_item.c 	if (atomic_read(&dqp->q_pincount) > 0) {
atomic_read        56 fs/xfs/xfs_extfree_item.c 	ASSERT(atomic_read(&efip->efi_refcount) > 0);
atomic_read       101 fs/xfs/xfs_extfree_item.c 	ASSERT(atomic_read(&efip->efi_next_extent) ==
atomic_read       756 fs/xfs/xfs_file.c 			atomic_read(&page->_refcount) == 1, TASK_INTERRUPTIBLE,
atomic_read        78 fs/xfs/xfs_filestream.c 	ret = atomic_read(&pag->pagf_fstrms);
atomic_read        55 fs/xfs/xfs_icache.c 	ASSERT(atomic_read(&ip->i_pincount) == 0);
atomic_read       115 fs/xfs/xfs_icache.c 	ASSERT(atomic_read(&ip->i_pincount) == 0);
atomic_read      1734 fs/xfs/xfs_icache.c 	    atomic_read(&VFS_I(ip)->i_dio_count))
atomic_read      1522 fs/xfs/xfs_inode.c 	ASSERT(!atomic_read(&VFS_I(ip)->i_count) ||
atomic_read       445 fs/xfs/xfs_inode.h #define xfs_ipincount(ip)	((unsigned int) atomic_read(&ip->i_pincount))
atomic_read       463 fs/xfs/xfs_inode_item.c 	ASSERT(atomic_read(&ip->i_pincount) > 0);
atomic_read      1870 fs/xfs/xfs_log.c 	ASSERT(atomic_read(&iclog->ic_refcnt) == 0);
atomic_read      2936 fs/xfs/xfs_log.c 	ASSERT(atomic_read(&iclog->ic_refcnt) == 0);
atomic_read      3184 fs/xfs/xfs_log.c 	ASSERT(atomic_read(&iclog->ic_refcnt) > 0);
atomic_read      3317 fs/xfs/xfs_log.c 	     atomic_read(&iclog->ic_refcnt) == 0 && iclog->ic_offset == 0)) {
atomic_read      3331 fs/xfs/xfs_log.c 		if (atomic_read(&iclog->ic_refcnt) == 0) {
atomic_read      3542 fs/xfs/xfs_log.c 	ASSERT(atomic_read(&ticket->t_ref) > 0);
atomic_read      3551 fs/xfs/xfs_log.c 	ASSERT(atomic_read(&ticket->t_ref) > 0);
atomic_read       129 fs/xfs/xfs_mount.c 	ASSERT(atomic_read(&pag->pag_ref) == 0);
atomic_read       148 fs/xfs/xfs_mount.c 		ASSERT(atomic_read(&pag->pag_ref) == 0);
atomic_read       117 fs/xfs/xfs_pwork.c 				atomic_read(&pctl->nr_work) == 0, HZ) == 0)
atomic_read       155 fs/xfs/xfs_qm.c 	ASSERT(atomic_read(&dqp->q_pincount) == 0);
atomic_read        51 fs/xfs/xfs_refcount_item.c 	ASSERT(atomic_read(&cuip->cui_refcount) > 0);
atomic_read        86 fs/xfs/xfs_refcount_item.c 	ASSERT(atomic_read(&cuip->cui_next_extent) ==
atomic_read        51 fs/xfs/xfs_rmap_item.c 	ASSERT(atomic_read(&ruip->rui_refcount) > 0);
atomic_read        85 fs/xfs/xfs_rmap_item.c 	ASSERT(atomic_read(&ruip->rui_next_extent) ==
atomic_read      1184 fs/xfs/xfs_super.c 	while (atomic_read(&mp->m_active_trans) > 0)
atomic_read      1203 fs/xfs/xfs_super.c 	WARN_ON(atomic_read(&mp->m_active_trans) != 0);
atomic_read       293 fs/xfs/xfs_trace.h 		__entry->hold = atomic_read(&bp->b_hold);
atomic_read       294 fs/xfs/xfs_trace.h 		__entry->pincount = atomic_read(&bp->b_pin_count);
atomic_read       362 fs/xfs/xfs_trace.h 		__entry->hold = atomic_read(&bp->b_hold);
atomic_read       363 fs/xfs/xfs_trace.h 		__entry->pincount = atomic_read(&bp->b_pin_count);
atomic_read       405 fs/xfs/xfs_trace.h 		__entry->hold = atomic_read(&bp->b_hold);
atomic_read       406 fs/xfs/xfs_trace.h 		__entry->pincount = atomic_read(&bp->b_pin_count);
atomic_read       445 fs/xfs/xfs_trace.h 		__entry->bli_refcount = atomic_read(&bip->bli_refcount);
atomic_read       449 fs/xfs/xfs_trace.h 		__entry->buf_hold = atomic_read(&bip->bli_buf->b_hold);
atomic_read       450 fs/xfs/xfs_trace.h 		__entry->buf_pincount = atomic_read(&bip->bli_buf->b_pin_count);
atomic_read       697 fs/xfs/xfs_trace.h 		__entry->count = atomic_read(&VFS_I(ip)->i_count);
atomic_read       698 fs/xfs/xfs_trace.h 		__entry->pincount = atomic_read(&ip->i_pincount);
atomic_read       146 fs/xfs/xfs_trans_buf.c 		ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       197 fs/xfs/xfs_trans_buf.c 		ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       294 fs/xfs/xfs_trans_buf.c 		ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       391 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       443 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       464 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       498 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       587 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       641 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       665 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       690 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       712 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read       745 fs/xfs/xfs_trans_buf.c 	ASSERT(atomic_read(&bip->bli_refcount) > 0);
atomic_read        29 include/asm-generic/atomic-instrumented.h #define atomic_read atomic_read
atomic_read       522 include/asm-generic/atomic-long.h 	return atomic_read(v);
atomic_read       170 include/asm-generic/atomic.h #ifndef atomic_read
atomic_read        42 include/asm-generic/qrwlock.h 	cnts = atomic_read(&lock->cnts);
atomic_read        61 include/asm-generic/qrwlock.h 	cnts = atomic_read(&lock->cnts);
atomic_read        26 include/asm-generic/qspinlock.h 	return atomic_read(&lock->val);
atomic_read        41 include/asm-generic/qspinlock.h 	return !atomic_read(&lock.val);
atomic_read        51 include/asm-generic/qspinlock.h 	return atomic_read(&lock->val) & ~_Q_LOCKED_MASK;
atomic_read        60 include/asm-generic/qspinlock.h 	u32 val = atomic_read(&lock->val);
atomic_read       211 include/crypto/if_alg.h 		     atomic_read(&ctx->rcvused), 0);
atomic_read        62 include/drm/spsc_queue.h 	return atomic_read(&queue->job_count);
atomic_read      1086 include/linux/atomic-fallback.h 	int c = atomic_read(v);
atomic_read      1136 include/linux/atomic-fallback.h 	int c = atomic_read(v);
atomic_read      1152 include/linux/atomic-fallback.h 	int c = atomic_read(v);
atomic_read      1168 include/linux/atomic-fallback.h 	int dec, c = atomic_read(v);
atomic_read       310 include/linux/blk-cgroup.h 		if (atomic_read(&css->cgroup->congestion_count)) {
atomic_read       773 include/linux/blk-cgroup.h 	int old = atomic_read(&blkg->use_delay);
atomic_read       801 include/linux/blk-cgroup.h 	int old = atomic_read(&blkg->use_delay);
atomic_read       661 include/linux/blkdev.h #define blk_queue_pm_only(q)	atomic_read(&(q)->pm_only)
atomic_read       211 include/linux/console.h 	WARN_ON(!atomic_read(&ignore_console_lock_warning) &&		\
atomic_read       112 include/linux/cpumask.h 	return atomic_read(&__num_online_cpus);
atomic_read       216 include/linux/firewire.h 	return atomic_read(&device->state) == FW_DEVICE_SHUTDOWN;
atomic_read        37 include/linux/freezer.h 	if (likely(!atomic_read(&system_freezing_cnt)))
atomic_read       563 include/linux/fs.h 	return atomic_read(&mapping->i_mmap_writable) > 0;
atomic_read      2822 include/linux/fs.h 	return atomic_read(&mapping->nr_thps);
atomic_read      2943 include/linux/fs.h 	return atomic_read(&inode->i_writecount) > 0;
atomic_read      2949 include/linux/fs.h 	BUG_ON(!atomic_read(&inode->i_readcount));
atomic_read       133 include/linux/iocontext.h 	WARN_ON_ONCE(atomic_read(&ioc->active_ref) <= 0);
atomic_read       142 include/linux/iocontext.h 	WARN_ON_ONCE(atomic_read(&ioc->nr_tasks) <= 0);
atomic_read      1588 include/linux/jbd2.h 		unsigned long committing = atomic_read(&journal->
atomic_read       254 include/linux/jump_label.h 	return atomic_read(&key->enabled);
atomic_read       308 include/linux/jump_label.h 	if (atomic_read(&key->enabled) != 0) {
atomic_read       309 include/linux/jump_label.h 		WARN_ON_ONCE(atomic_read(&key->enabled) != 1);
atomic_read       319 include/linux/jump_label.h 	if (atomic_read(&key->enabled) != 1) {
atomic_read       320 include/linux/jump_label.h 		WARN_ON_ONCE(atomic_read(&key->enabled) != 0);
atomic_read       326 include/linux/kgdb.h 	(raw_smp_processor_id() == atomic_read(&kgdb_active))
atomic_read       541 include/linux/kvm_host.h 	int num_vcpus = atomic_read(&kvm->online_vcpus);
atomic_read       551 include/linux/kvm_host.h 	     idx < atomic_read(&kvm->online_vcpus) && \
atomic_read      1014 include/linux/kvm_host.h 	int slot = atomic_read(&slots->lru_slot);
atomic_read       708 include/linux/mm.h 	return atomic_read(compound_mapcount_ptr(page)) + 1;
atomic_read       735 include/linux/mm.h 	return atomic_read(&page->_mapcount) + 1;
atomic_read       628 include/linux/mm_types.h 	return atomic_read(&mm->tlb_flush_pending);
atomic_read       640 include/linux/mm_types.h 	return atomic_read(&mm->tlb_flush_pending) > 1;
atomic_read        38 include/linux/osq_lock.h 	return atomic_read(&lock->tail) != OSQ_UNLOCKED_VAL;
atomic_read       641 include/linux/page-flags.h 		return atomic_read(&page->_mapcount) < 0;
atomic_read       645 include/linux/page-flags.h 	return atomic_read(&page->_mapcount) ==
atomic_read       646 include/linux/page-flags.h 	       atomic_read(compound_mapcount_ptr(head));
atomic_read        67 include/linux/page_ref.h 	return atomic_read(&page->_refcount);
atomic_read        72 include/linux/page_ref.h 	return atomic_read(&compound_head(page)->_refcount);
atomic_read      1152 include/linux/pci.h 	return (atomic_read(&pdev->enable_cnt) > 0);
atomic_read        52 include/linux/quotaops.h 	WARN_ON_ONCE(!atomic_read(&dquot->dq_count));
atomic_read        62 include/linux/quotaops.h 	if (atomic_read(&dquot->dq_count) > 1)
atomic_read        43 include/linux/refcount.h 	return atomic_read(&r->refs);
atomic_read       186 include/linux/rhashtable.h 	return atomic_read(&ht->nelems) > (tbl->size / 4 * 3) &&
atomic_read       199 include/linux/rhashtable.h 	return atomic_read(&ht->nelems) < (tbl->size * 3 / 10) &&
atomic_read       211 include/linux/rhashtable.h 	return atomic_read(&ht->nelems) > tbl->size &&
atomic_read       223 include/linux/rhashtable.h 	return atomic_read(&ht->nelems) >= ht->max_elems;
atomic_read       300 include/linux/rio.h 	return atomic_read(&mport->state) == RIO_DEVICE_RUNNING;
atomic_read       506 include/linux/sbitmap.h 	int old = atomic_read(index);
atomic_read       522 include/linux/sbitmap.h 	ws = &sbq->ws[atomic_read(wait_index)];
atomic_read       369 include/linux/sched/mm.h 	if (likely(!(atomic_read(&mm->membarrier_state) &
atomic_read      1603 include/linux/skbuff.h 	       (atomic_read(&skb_shinfo(skb)->dataref) & SKB_DATAREF_MASK) != 1;
atomic_read      1630 include/linux/skbuff.h 	dataref = atomic_read(&skb_shinfo(skb)->dataref);
atomic_read       112 include/linux/sysctl.h 	return (void *)(unsigned long)atomic_read(&poll->event);
atomic_read        22 include/misc/cxl-base.h        return (atomic_read(&cxl_use_count) != 0);
atomic_read       976 include/net/bluetooth/hci_core.h 	BT_DBG("hcon %p orig refcnt %d", conn, atomic_read(&conn->refcnt));
atomic_read       984 include/net/bluetooth/hci_core.h 	BT_DBG("hcon %p orig refcnt %d", conn, atomic_read(&conn->refcnt));
atomic_read       807 include/net/bluetooth/l2cap.h 	mutex_lock_nested(&chan->lock, atomic_read(&chan->nesting));
atomic_read       105 include/net/bonding.h 		return atomic_read(&netpoll_block_tx);
atomic_read       190 include/net/dn_nsp.h         return atomic_read(&sk->sk_rmem_alloc) > (sk->sk_rcvbuf >> 1);
atomic_read      1266 include/net/ip_vs.h 	if (atomic_read(&ctl_cp->n_control) == 0) {
atomic_read      1508 include/net/ip_vs.h #define IP_VS_DFWD_METHOD(dest) (atomic_read(&(dest)->conn_flags) & \
atomic_read      1682 include/net/ip_vs.h 	return (atomic_read(&dest->activeconns) << 8) +
atomic_read      1683 include/net/ip_vs.h 		atomic_read(&dest->inactconns);
atomic_read       221 include/net/llc_c_ev.h 	return atomic_read(&sk->sk_rmem_alloc) + skb->truesize <
atomic_read       428 include/net/net_namespace.h 	return atomic_read(&net->ipv4.rt_genid);
atomic_read       434 include/net/net_namespace.h 	return atomic_read(&net->ipv6.fib6_sernum);
atomic_read       467 include/net/net_namespace.h 	return atomic_read(&net->fnhe_genid);
atomic_read       224 include/net/request_sock.h 	return atomic_read(&queue->qlen);
atomic_read       229 include/net/request_sock.h 	return atomic_read(&queue->young);
atomic_read       923 include/net/sock.h 	unsigned int qsize = sk->sk_backlog.len + atomic_read(&sk->sk_rmem_alloc);
atomic_read      1203 include/net/sock.h 	       sk->sk_prot->name, sk, atomic_read(&sk->sk_prot->socks));
atomic_read      2051 include/net/sock.h 	return atomic_read(&sk->sk_rmem_alloc);
atomic_read      2327 include/net/sock.h 						atomic_read(&sk->sk_drops) : 0;
atomic_read       689 include/net/tcp.h 	    atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf &&
atomic_read      1396 include/net/tcp.h 				  atomic_read(&sk->sk_rmem_alloc));
atomic_read      1414 include/net/tcp.h 	return atomic_read(&sk->sk_rmem_alloc) > threshold;
atomic_read      1329 include/net/xfrm.h 	return atomic_read(&x->tunnel_users);
atomic_read      2878 include/rdma/ib_verbs.h 	if (atomic_read(usecnt) && ib_is_destroy_retryable(-EBUSY, why, uobj))
atomic_read       195 include/sound/hdaudio.h 	return atomic_read(&codec->in_pm);
atomic_read       481 include/trace/events/bcache.h 		__entry->blocked	= atomic_read(&ca->set->prio_blocked);
atomic_read      1003 include/trace/events/btrfs.h 		__entry->refs		= atomic_read(&buf->refs);
atomic_read       190 include/trace/events/filelock.h 		__entry->wcount = atomic_read(&inode->i_writecount);
atomic_read       191 include/trace/events/filelock.h 		__entry->rcount = atomic_read(&inode->i_readcount);
atomic_read       192 include/trace/events/filelock.h 		__entry->icount = atomic_read(&inode->i_count);
atomic_read       184 include/trace/events/fscache.h 		    __entry->n_children	= atomic_read(&cookie->n_children);
atomic_read       185 include/trace/events/fscache.h 		    __entry->n_active	= atomic_read(&cookie->n_active);
atomic_read       233 include/trace/events/fscache.h 		    __entry->p_usage		= atomic_read(&cookie->parent->usage);
atomic_read       234 include/trace/events/fscache.h 		    __entry->p_n_children	= atomic_read(&cookie->parent->n_children);
atomic_read       263 include/trace/events/fscache.h 		    __entry->usage	= atomic_read(&cookie->usage);
atomic_read       264 include/trace/events/fscache.h 		    __entry->n_children	= atomic_read(&cookie->n_children);
atomic_read       265 include/trace/events/fscache.h 		    __entry->n_active	= atomic_read(&cookie->n_active);
atomic_read       291 include/trace/events/fscache.h 		    __entry->usage	= atomic_read(&cookie->usage);
atomic_read       292 include/trace/events/fscache.h 		    __entry->n_children	= atomic_read(&cookie->n_children);
atomic_read       293 include/trace/events/fscache.h 		    __entry->n_active	= atomic_read(&cookie->n_active);
atomic_read       317 include/trace/events/fscache.h 		    __entry->usage	= atomic_read(&cookie->usage);
atomic_read       318 include/trace/events/fscache.h 		    __entry->n_children	= atomic_read(&cookie->n_children);
atomic_read       319 include/trace/events/fscache.h 		    __entry->n_active	= atomic_read(&cookie->n_active);
atomic_read        84 include/trace/events/module.h 		__entry->refcnt	= atomic_read(&mod->refcnt);
atomic_read        47 include/trace/events/neigh.h 		__entry->entries = atomic_read(&tbl->gc_entries);
atomic_read      1818 include/trace/events/rpcrdma.h 		__entry->avail = atomic_read(&rdma->sc_sq_avail);
atomic_read      1854 include/trace/events/rpcrdma.h 		__entry->avail = atomic_read(&rdma->sc_sq_avail);
atomic_read        38 include/trace/events/rpm.h 		__entry->usage_count = atomic_read(
atomic_read        44 include/trace/events/rpm.h 		__entry->child_count = atomic_read(
atomic_read       593 include/trace/events/rxrpc.h 		    __entry->usage = atomic_read(&conn->usage);
atomic_read        83 include/trace/events/sock.h 		__entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);
atomic_read       115 include/trace/events/sock.h 		__entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);
atomic_read        29 include/trace/events/vb2.h 			atomic_read(&q->owned_by_drv_count);
atomic_read       479 ipc/msg.c      		msginfo->msgmap = atomic_read(&ns->msg_hdrs);
atomic_read       480 ipc/msg.c      		msginfo->msgtql = atomic_read(&ns->msg_bytes);
atomic_read       179 kernel/async.c 	if (!entry || atomic_read(&entry_count) > MAX_WORK) {
atomic_read       373 kernel/audit.c 				atomic_read(&audit_lost),
atomic_read      1202 kernel/audit.c 		s.lost			= atomic_read(&audit_lost);
atomic_read       674 kernel/cgroup/cgroup-v1.c 			   atomic_read(&ss->root->nr_cgrps),
atomic_read      1327 kernel/cgroup/cgroup.c 	BUG_ON(atomic_read(&root->nr_cgrps));
atomic_read      2073 kernel/cgroup/cgroup.c 	BUG_ON(atomic_read(&root->nr_cgrps) != 1);
atomic_read      4558 kernel/cgroup/cgroup.c 		    !atomic_read(&task->signal->live))
atomic_read       229 kernel/cgroup/debug.c 			  atomic_read(&css->online_cnt), pbuf);
atomic_read        75 kernel/cred.c  	return atomic_read(&cred->subscribers);
atomic_read       101 kernel/cred.c  	    atomic_read(&cred->usage) != 0 ||
atomic_read       106 kernel/cred.c  		      atomic_read(&cred->usage),
atomic_read       109 kernel/cred.c  	if (atomic_read(&cred->usage) != 0)
atomic_read       111 kernel/cred.c  		      cred, atomic_read(&cred->usage));
atomic_read       135 kernel/cred.c  	       atomic_read(&cred->usage),
atomic_read       138 kernel/cred.c  	BUG_ON(atomic_read(&cred->usage) != 0);
atomic_read       162 kernel/cred.c  	       atomic_read(&tsk->cred->usage),
atomic_read       349 kernel/cred.c  		       p->cred, atomic_read(&p->cred->usage),
atomic_read       440 kernel/cred.c  	       atomic_read(&new->usage),
atomic_read       449 kernel/cred.c  	BUG_ON(atomic_read(&new->usage) < 1);
atomic_read       523 kernel/cred.c  	       atomic_read(&new->usage),
atomic_read       529 kernel/cred.c  	BUG_ON(atomic_read(&new->usage) < 1);
atomic_read       546 kernel/cred.c  	       atomic_read(&new->usage),
atomic_read       569 kernel/cred.c  	       atomic_read(&old->usage),
atomic_read       587 kernel/cred.c  	       atomic_read(&old->usage),
atomic_read       812 kernel/cred.c  	       atomic_read(&cred->usage),
atomic_read       886 kernel/cred.c  	       atomic_read(&tsk->cred->usage),
atomic_read       459 kernel/debug/debug_core.c 	if (atomic_read(&kgdb_setting_breakpoint))
atomic_read       476 kernel/debug/debug_core.c 	if (atomic_read(&kgdb_active) != raw_smp_processor_id())
atomic_read       615 kernel/debug/debug_core.c 	if (atomic_read(&kgdb_cpu_doing_single_step) != -1 &&
atomic_read       663 kernel/debug/debug_core.c 	       (atomic_read(&masters_in_kgdb) + atomic_read(&slaves_in_kgdb)) !=
atomic_read       712 kernel/debug/debug_core.c 		while (kgdb_do_roundup && atomic_read(&slaves_in_kgdb))
atomic_read       717 kernel/debug/debug_core.c 	if (atomic_read(&kgdb_cpu_doing_single_step) != -1) {
atomic_read       718 kernel/debug/debug_core.c 		int sstep_cpu = atomic_read(&kgdb_cpu_doing_single_step);
atomic_read       855 kernel/debug/debug_core.c 	if (!kgdb_connected || atomic_read(&kgdb_active) != -1 || dbg_kdb_mode)
atomic_read      1011 kernel/debug/debug_core.c 	if (atomic_read(&kgdb_break_tasklet_var) ||
atomic_read      1012 kernel/debug/debug_core.c 		atomic_read(&kgdb_active) != -1 ||
atomic_read      1013 kernel/debug/debug_core.c 		atomic_read(&kgdb_setting_breakpoint))
atomic_read       439 kernel/debug/gdbstub.c 		tid = -atomic_read(&kgdb_active) - 2;
atomic_read        39 kernel/debug/kdb/kdb_debugger.c 	kdb_initial_cpu = atomic_read(&kgdb_active);
atomic_read        69 kernel/debug/kdb/kdb_debugger.c 	if (atomic_read(&kgdb_setting_breakpoint))
atomic_read      2181 kernel/debug/kdb/kdb_main.c 	if (atomic_read(&kdb_nmi_disabled))
atomic_read       251 kernel/events/callchain.c 	if (atomic_read(&nr_callchain_events))
atomic_read      3326 kernel/events/core.c 	if (atomic_read(&nr_switch_events))
atomic_read      3337 kernel/events/core.c 	if (atomic_read(this_cpu_ptr(&perf_cgroup_events)))
atomic_read      3576 kernel/events/core.c 	if (atomic_read(this_cpu_ptr(&perf_cgroup_events)))
atomic_read      3587 kernel/events/core.c 	if (atomic_read(&nr_switch_events))
atomic_read      5634 kernel/events/core.c 	if (atomic_read(&rb->mmap_count))
atomic_read      7089 kernel/events/core.c 	if (!atomic_read(&nr_comm_events) &&
atomic_read      7090 kernel/events/core.c 	    !atomic_read(&nr_mmap_events) &&
atomic_read      7091 kernel/events/core.c 	    !atomic_read(&nr_task_events))
atomic_read      7200 kernel/events/core.c 	if (!atomic_read(&nr_comm_events))
atomic_read      7298 kernel/events/core.c 	if (!atomic_read(&nr_namespaces_events))
atomic_read      7667 kernel/events/core.c 	if (!atomic_read(&nr_mmap_events))
atomic_read      7948 kernel/events/core.c 	if (!atomic_read(&nr_ksymbol_events))
atomic_read      8067 kernel/events/core.c 		if (atomic_read(&nr_ksymbol_events))
atomic_read      8074 kernel/events/core.c 	if (!atomic_read(&nr_bpf_events))
atomic_read      8186 kernel/events/core.c 	int events = atomic_read(&event->event_limit);
atomic_read      10393 kernel/events/core.c 		if (!atomic_read(&perf_sched_count)) {
atomic_read      10794 kernel/events/core.c 	if (atomic_read(&event->mmap_count))
atomic_read       390 kernel/events/ring_buffer.c 	if (!atomic_read(&rb->aux_mmap_count))
atomic_read      1436 kernel/events/uprobes.c 	if (!atomic_read(&vma->vm_mm->mm_users)) /* called by mmput() ? */
atomic_read      1604 kernel/events/uprobes.c 		wait_event(area->wq, (atomic_read(&area->slot_count) < UINSNS_PER_PAGE));
atomic_read       363 kernel/exit.c  	if (atomic_read(&mm->mm_users) <= 1) {
atomic_read      1073 kernel/fork.c  	VM_BUG_ON(atomic_read(&mm->mm_users));
atomic_read      1300 kernel/fork.c  		    atomic_read(&mm->mm_users) > 1) {
atomic_read      1872 kernel/fork.c  	if (atomic_read(&p->real_cred->user->processes) >=
atomic_read      2792 kernel/fork.c  	    (fd && atomic_read(&fd->count) > 1)) {
atomic_read       373 kernel/futex.c 	return atomic_read(&hb->waiters);
atomic_read       104 kernel/irq/manage.c 		return !atomic_read(&desc->threads_active);
atomic_read       138 kernel/irq/manage.c 			   !atomic_read(&desc->threads_active));
atomic_read       337 kernel/irq/spurious.c 			handled = atomic_read(&desc->threads_handled);
atomic_read       110 kernel/jump_label.c 	int n = atomic_read(&key->enabled);
atomic_read       135 kernel/jump_label.c 	for (v = atomic_read(&key->enabled); v > 0; v = v1) {
atomic_read       142 kernel/jump_label.c 	if (atomic_read(&key->enabled) == 0) {
atomic_read       169 kernel/jump_label.c 	if (atomic_read(&key->enabled) > 0) {
atomic_read       170 kernel/jump_label.c 		WARN_ON_ONCE(atomic_read(&key->enabled) != 1);
atomic_read       175 kernel/jump_label.c 	if (atomic_read(&key->enabled) == 0) {
atomic_read       200 kernel/jump_label.c 	if (atomic_read(&key->enabled) != 1) {
atomic_read       201 kernel/jump_label.c 		WARN_ON_ONCE(atomic_read(&key->enabled) != 0);
atomic_read       154 kernel/kmod.c  				    atomic_read(&kmod_concurrent_max),
atomic_read       821 kernel/locking/locktorture.c 	if (atomic_read(&cxt.n_lock_torture_errors))
atomic_read        58 kernel/locking/osq_lock.c 		if (atomic_read(&lock->tail) == curr &&
atomic_read        68 kernel/locking/qrwlock.c 	if (!atomic_read(&lock->cnts) &&
atomic_read       221 kernel/locking/qspinlock.c 	u32 old, new, val = atomic_read(&lock->val);
atomic_read        88 kernel/locking/qspinlock_paravirt.h 		int val = atomic_read(&lock->val);
atomic_read       133 kernel/locking/qspinlock_paravirt.h 	int val = atomic_read(&lock->val);
atomic_read       485 kernel/locking/qspinlock_paravirt.h 	return (u32)(atomic_read(&lock->val) | _Q_LOCKED_VAL);
atomic_read       500 kernel/locking/qspinlock_paravirt.h 		     (unsigned long)lock, atomic_read(&lock->val));
atomic_read       963 kernel/module.c 	return atomic_read(&mod->refcnt) - MODULE_REF_BASE;
atomic_read       137 kernel/padata.c 	if (atomic_read(&pd->refcnt) >= MAX_OBJ_NUM)
atomic_read       297 kernel/power/swap.c 	wait_event(hb->wait, atomic_read(&hb->count) == 0);
atomic_read       597 kernel/power/swap.c 		wait_event(d->go, atomic_read(&d->ready) ||
atomic_read       640 kernel/power/swap.c 		wait_event(d->go, atomic_read(&d->ready) ||
atomic_read       805 kernel/power/swap.c 			           atomic_read(&data[thr].stop));
atomic_read       844 kernel/power/swap.c 		wait_event(crc->done, atomic_read(&crc->stop));
atomic_read      1124 kernel/power/swap.c 		wait_event(d->go, atomic_read(&d->ready) ||
atomic_read      1330 kernel/power/swap.c 			wait_event(crc->done, atomic_read(&crc->stop));
atomic_read      1385 kernel/power/swap.c 			           atomic_read(&data[thr].stop));
atomic_read      1430 kernel/power/swap.c 		wait_event(crc->done, atomic_read(&crc->stop));
atomic_read        75 kernel/printk/printk_safe.c 	len = atomic_read(&s->len);
atomic_read       200 kernel/printk/printk_safe.c 	len = atomic_read(&s->len);
atomic_read       266 kernel/rcu/rcu.h 	if (!atomic_read(&___rfd_beenhere) && \
atomic_read       311 kernel/rcu/rcuperf.c 	if (atomic_read(&n_rcu_perf_writer_finished) < nrealwriters)
atomic_read       406 kernel/rcu/rcuperf.c 			if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) {
atomic_read       430 kernel/rcu/rcuperf.c 		    atomic_read(&n_rcu_perf_writer_started) >= nrealwriters)
atomic_read       459 kernel/rcu/rcuperf.c 		    atomic_read(&n_rcu_perf_writer_finished) >= nrealwriters)
atomic_read       594 kernel/rcu/rcuperf.c 			   atomic_read(&n_rcu_perf_writer_finished) >=
atomic_read       596 kernel/rcu/rcuperf.c 	} while (atomic_read(&n_rcu_perf_writer_finished) < nrealwriters);
atomic_read       665 kernel/rcu/rcuperf.c 	while (atomic_read(&n_rcu_perf_reader_started) < nrealreaders)
atomic_read      1426 kernel/rcu/rcutorture.c 		atomic_read(&n_rcu_torture_alloc),
atomic_read      1427 kernel/rcu/rcutorture.c 		atomic_read(&n_rcu_torture_alloc_fail),
atomic_read      1428 kernel/rcu/rcutorture.c 		atomic_read(&n_rcu_torture_free));
atomic_read      1430 kernel/rcu/rcutorture.c 		atomic_read(&n_rcu_torture_mberror),
atomic_read      1445 kernel/rcu/rcutorture.c 	if (atomic_read(&n_rcu_torture_mberror) != 0 ||
atomic_read      1469 kernel/rcu/rcutorture.c 		pr_cont(" %d", atomic_read(&rcu_torture_wcount[i]));
atomic_read      2058 kernel/rcu/rcutorture.c 			   atomic_read(&barrier_cbs_count) == 0 ||
atomic_read      2064 kernel/rcu/rcutorture.c 		if (atomic_read(&barrier_cbs_invoked) != n_barrier_cbs) {
atomic_read      2067 kernel/rcu/rcutorture.c 			       atomic_read(&barrier_cbs_invoked),
atomic_read      2249 kernel/rcu/rcutorture.c 	if (atomic_read(&n_rcu_torture_error) || n_rcu_torture_barrier_error)
atomic_read       287 kernel/rcu/tree.c 	if (atomic_read(&rdp->dynticks) & RCU_DYNTICK_CTRL_CTR)
atomic_read       301 kernel/rcu/tree.c 	return !(atomic_read(&rdp->dynticks) & RCU_DYNTICK_CTRL_CTR);
atomic_read       348 kernel/rcu/tree.c 		old = atomic_read(&rdp->dynticks);
atomic_read       579 kernel/rcu/tree.c 	trace_rcu_dyntick(TPS("Start"), rdp->dynticks_nesting, 0, atomic_read(&rdp->dynticks));
atomic_read       653 kernel/rcu/tree.c 				  atomic_read(&rdp->dynticks));
atomic_read       660 kernel/rcu/tree.c 	trace_rcu_dyntick(TPS("Startirq"), rdp->dynticks_nmi_nesting, 0, atomic_read(&rdp->dynticks));
atomic_read       747 kernel/rcu/tree.c 	trace_rcu_dyntick(TPS("End"), rdp->dynticks_nesting, 1, atomic_read(&rdp->dynticks));
atomic_read       831 kernel/rcu/tree.c 			  rdp->dynticks_nmi_nesting + incby, atomic_read(&rdp->dynticks));
atomic_read      2838 kernel/rcu/tree.c 			  atomic_read(&rcu_state.barrier_cpu_count), done);
atomic_read      1539 kernel/rcu/tree_plugin.h 	while (WARN_ON_ONCE(atomic_read(&rdp->nocb_lock_contended)))
atomic_read      2427 kernel/rcu/tree_plugin.h 		"cC"[!!atomic_read(&rdp->nocb_lock_contended)],
atomic_read       619 kernel/rcu/tree_stall.h 	    atomic_read(&warned))
atomic_read       628 kernel/rcu/tree_stall.h 	    atomic_read(&warned)) {
atomic_read       155 kernel/rcu/update.c 	return rcu_expedited || atomic_read(&rcu_expedited_nesting);
atomic_read       638 kernel/sched/core.c 	if (!(atomic_read(nohz_flags(cpu)) & NOHZ_KICK_MASK))
atomic_read      3441 kernel/sched/core.c 	return atomic_read(&cpu_rq(cpu)->nr_iowait);
atomic_read        71 kernel/sched/cpupri.c 		if (!atomic_read(&(vec)->count))
atomic_read       223 kernel/sched/cputime.c 	if (atomic_read(&rq->nr_iowait) > 0)
atomic_read       380 kernel/sched/deadline.c 	return atomic_read(&rq->rd->dlo_count);
atomic_read      2632 kernel/sched/fair.c 		mm_users = atomic_read(&mm->mm_users);
atomic_read      9421 kernel/sched/fair.c 	if (likely(!atomic_read(&nohz.nr_cpus)))
atomic_read      9498 kernel/sched/fair.c 		nr_busy = atomic_read(&sds->nr_busy_cpus);
atomic_read      9734 kernel/sched/fair.c 	if (!(atomic_read(nohz_flags(this_cpu)) & NOHZ_KICK_MASK))
atomic_read        40 kernel/sched/membarrier.c 		       atomic_read(&mm->membarrier_state));
atomic_read       141 kernel/sched/membarrier.c 		if (!(atomic_read(&mm->membarrier_state) &
atomic_read       145 kernel/sched/membarrier.c 		if (!(atomic_read(&mm->membarrier_state) &
atomic_read       150 kernel/sched/membarrier.c 	if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1)
atomic_read       202 kernel/sched/membarrier.c 	int membarrier_state = atomic_read(&mm->membarrier_state);
atomic_read       206 kernel/sched/membarrier.c 	if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) {
atomic_read       265 kernel/sched/membarrier.c 	if (atomic_read(&mm->membarrier_state) &
atomic_read       298 kernel/sched/membarrier.c 	if ((atomic_read(&mm->membarrier_state) & ready_state) == ready_state)
atomic_read       271 kernel/sched/rt.c 	return atomic_read(&rq->rd->rto_count);
atomic_read      2492 kernel/sched/sched.h 	membarrier_state = atomic_read(&next_mm->membarrier_state);
atomic_read      1225 kernel/sched/topology.c 		if (!atomic_read(&d->rd->refcount))
atomic_read      1268 kernel/sched/topology.c 	if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref))
atomic_read      1271 kernel/sched/topology.c 	if (atomic_read(&(*per_cpu_ptr(sdd->sg, cpu))->ref))
atomic_read      1274 kernel/sched/topology.c 	if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref))
atomic_read       337 kernel/smpboot.c 	return atomic_read(&per_cpu(cpu_hotplug_state, cpu));
atomic_read       359 kernel/smpboot.c 	switch (atomic_read(&per_cpu(cpu_hotplug_state, cpu))) {
atomic_read       428 kernel/smpboot.c 	if (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) == CPU_DEAD)
atomic_read       433 kernel/smpboot.c 	while (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) != CPU_DEAD) {
atomic_read       441 kernel/smpboot.c 	oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu));
atomic_read       472 kernel/smpboot.c 		oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu));
atomic_read       519 kernel/softirq.c 			if (!atomic_read(&t->count)) {
atomic_read       467 kernel/sys.c   	if (atomic_read(&new_user->processes) >= rlimit(RLIMIT_NPROC) &&
atomic_read       198 kernel/time/clocksource.c 	reset_pending = atomic_read(&watchdog_reset_pending);
atomic_read       216 kernel/time/clocksource.c 		    atomic_read(&watchdog_reset_pending)) {
atomic_read       234 kernel/time/clocksource.c 		if (atomic_read(&watchdog_reset_pending))
atomic_read      1204 kernel/time/hrtimer.c 	if (atomic_read(&cpu_base->timer_waiters)) {
atomic_read       183 kernel/time/tick-sched.c 	int val = atomic_read(dep);
atomic_read       389 kernel/time/tick-sched.c 		if (atomic_read(&current->tick_dep_mask) ||
atomic_read       390 kernel/time/tick-sched.c 		    atomic_read(&current->signal->tick_dep_mask))
atomic_read      1259 kernel/time/timer.c 	if (atomic_read(&base->timer_waiters)) {
atomic_read       376 kernel/trace/blktrace.c 	snprintf(buf, sizeof(buf), "%u\n", atomic_read(&bt->dropped));
atomic_read       170 kernel/trace/fgraph.c 	trace->overrun = atomic_read(&current->trace_overrun);
atomic_read      1746 kernel/trace/ring_buffer.c 	if (atomic_read(&buffer->resize_disabled))
atomic_read      1854 kernel/trace/ring_buffer.c 	if (atomic_read(&buffer->record_disabled)) {
atomic_read      2973 kernel/trace/ring_buffer.c 	if (unlikely(atomic_read(&buffer->record_disabled)))
atomic_read      2983 kernel/trace/ring_buffer.c 	if (unlikely(atomic_read(&cpu_buffer->record_disabled)))
atomic_read      3128 kernel/trace/ring_buffer.c 	if (atomic_read(&buffer->record_disabled))
atomic_read      3138 kernel/trace/ring_buffer.c 	if (atomic_read(&cpu_buffer->record_disabled))
atomic_read      3232 kernel/trace/ring_buffer.c 		rd = atomic_read(&buffer->record_disabled);
atomic_read      3255 kernel/trace/ring_buffer.c 		rd = atomic_read(&buffer->record_disabled);
atomic_read      3269 kernel/trace/ring_buffer.c 	return !atomic_read(&buffer->record_disabled);
atomic_read      3285 kernel/trace/ring_buffer.c 	return !(atomic_read(&buffer->record_disabled) & RB_BUFFER_OFF);
atomic_read      4533 kernel/trace/ring_buffer.c 	if (atomic_read(&buffer_a->record_disabled))
atomic_read      4536 kernel/trace/ring_buffer.c 	if (atomic_read(&buffer_b->record_disabled))
atomic_read      4539 kernel/trace/ring_buffer.c 	if (atomic_read(&cpu_buffer_a->record_disabled))
atomic_read      4542 kernel/trace/ring_buffer.c 	if (atomic_read(&cpu_buffer_b->record_disabled))
atomic_read      2249 kernel/trace/trace.c 	if (atomic_read(&trace_record_taskinfo_disabled) || !tracing_is_on())
atomic_read        59 kernel/trace/trace_branch.c 	if (atomic_read(&data->disabled))
atomic_read      1151 kernel/trace/trace_events_hist.c 	if (unlikely(atomic_read(&tp->key.enabled) > 0)) {
atomic_read       147 kernel/trace/trace_functions.c 	if (!atomic_read(&data->disabled)) {
atomic_read       387 kernel/trace/trace_irqsoff.c 	if (unlikely(!data) || atomic_read(&data->disabled))
atomic_read       426 kernel/trace/trace_irqsoff.c 	    !data->critical_start || atomic_read(&data->disabled))
atomic_read       178 kernel/ucount.c 	c = atomic_read(v);
atomic_read       335 kernel/umh.c   					atomic_read(&running_helpers) == 0,
atomic_read       770 kernel/workqueue.c 	return !atomic_read(&pool->nr_running);
atomic_read       796 kernel/workqueue.c 		atomic_read(&pool->nr_running) <= 1;
atomic_read      1796 kernel/workqueue.c 		     atomic_read(&pool->nr_running));
atomic_read      2731 kernel/workqueue.c 		WARN_ON_ONCE(atomic_read(&wq->nr_pwqs_to_flush));
atomic_read       114 lib/dump_stack.c 		do { cpu_relax(); } while (atomic_read(&dump_lock) != -1);
atomic_read        49 lib/fault-inject.c 		       atomic_read(&attr->space),
atomic_read        50 lib/fault-inject.c 		       atomic_read(&attr->times));
atomic_read       123 lib/fault-inject.c 	if (atomic_read(&attr->times) == 0)
atomic_read       126 lib/fault-inject.c 	if (atomic_read(&attr->space) > size) {
atomic_read       146 lib/fault-inject.c 	if (atomic_read(&attr->times) != -1)
atomic_read        22 lib/is_single_threaded.c 	if (atomic_read(&task->signal->live) != 1)
atomic_read        25 lib/is_single_threaded.c 	if (atomic_read(&mm->mm_users) == 1)
atomic_read        66 lib/refcount.c 	unsigned int new, val = atomic_read(&r->refs);
atomic_read       123 lib/refcount.c 	unsigned int new, val = atomic_read(&r->refs);
atomic_read       182 lib/refcount.c 	unsigned int new, val = atomic_read(&r->refs);
atomic_read       277 lib/refcount.c 	unsigned int new, val = atomic_read(&r->refs);
atomic_read       381 lib/rhashtable.c 	unsigned int nelems = atomic_read(&ht->nelems);
atomic_read       509 lib/sbitmap.c  	if (!atomic_read(&sbq->ws_active))
atomic_read       512 lib/sbitmap.c  	wake_index = atomic_read(&sbq->wake_index);
atomic_read       517 lib/sbitmap.c  			if (wake_index != atomic_read(&sbq->wake_index))
atomic_read       615 lib/sbitmap.c  	wake_index = atomic_read(&sbq->wake_index);
atomic_read       645 lib/sbitmap.c  	seq_printf(m, "wake_index=%d\n", atomic_read(&sbq->wake_index));
atomic_read       646 lib/sbitmap.c  	seq_printf(m, "ws_active=%d\n", atomic_read(&sbq->ws_active));
atomic_read       653 lib/sbitmap.c  			   atomic_read(&ws->wait_cnt),
atomic_read       202 lib/test_rhashtable.c 		total, atomic_read(&ht->nelems), entries, chain_len);
atomic_read       204 lib/test_rhashtable.c 	if (total != atomic_read(&ht->nelems) || total != entries)
atomic_read       638 lib/test_rhashtable.c 	if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == -1)) {
atomic_read       793 lib/test_rhashtable.c 	if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == 0))
atomic_read      1136 mm/backing-dev.c 	if (atomic_read(&nr_wb_congested[sync]) == 0) {
atomic_read       171 mm/debug.c     		mm->pgd, atomic_read(&mm->mm_users),
atomic_read       172 mm/debug.c     		atomic_read(&mm->mm_count),
atomic_read       195 mm/debug.c     		atomic_read(&mm->tlb_flush_pending),
atomic_read       378 mm/frontswap.c 		totalpages += atomic_read(&si->frontswap_pages);
atomic_read       393 mm/frontswap.c 		si_frontswap_pages = atomic_read(&si->frontswap_pages);
atomic_read       140 mm/huge_memory.c 	return atomic_read(&huge_zero_refcount) == 1 ? HPAGE_PMD_NR : 0;
atomic_read      2441 mm/huge_memory.c 	VM_BUG_ON_PAGE(atomic_read(&page_tail->_mapcount) != -1, page_tail);
atomic_read      2586 mm/huge_memory.c 		return atomic_read(&page->_mapcount) + 1;
atomic_read      2593 mm/huge_memory.c 		ret += atomic_read(&page[i]._mapcount) + 1;
atomic_read      2634 mm/huge_memory.c 		mapcount = atomic_read(&page->_mapcount) + 1;
atomic_read      2644 mm/huge_memory.c 		mapcount = atomic_read(&page[i]._mapcount) + 1;
atomic_read       404 mm/khugepaged.c 	return atomic_read(&mm->mm_users) == 0;
atomic_read       664 mm/kmemleak.c  	WARN_ON(atomic_read(&object->use_count) < 1);
atomic_read      1416 mm/kmemleak.c  		if (atomic_read(&object->use_count) > 1) {
atomic_read      1418 mm/kmemleak.c  				 atomic_read(&object->use_count));
atomic_read       452 mm/ksm.c       	return atomic_read(&mm->mm_users) == 0;
atomic_read      1653 mm/memcontrol.c 	if (!atomic_read(&memcg->numainfo_events))
atomic_read      2111 mm/memcontrol.c 	if (atomic_read(&memcg->moving_account) <= 0)
atomic_read      4529 mm/memcontrol.c 		    atomic_read(&frn->done.cnt) == 1) {
atomic_read      4576 mm/memcontrol.c 		    atomic_read(&frn->done.cnt) == 1) {
atomic_read       765 mm/migrate.c   			if (atomic_read(&bh->b_count)) {
atomic_read       255 mm/mmu_notifier.c 	BUG_ON(atomic_read(&mm->mm_users) <= 0);
atomic_read       305 mm/mmu_notifier.c 	BUG_ON(atomic_read(&mm->mm_users) <= 0);
atomic_read       430 mm/mmu_notifier.c 	BUG_ON(atomic_read(&mm->mm_count) <= 0);
atomic_read       463 mm/mmu_notifier.c 	BUG_ON(atomic_read(&mm->mm_count) <= 0);
atomic_read        65 mm/mprotect.c  	    atomic_read(&vma->vm_mm->mm_users) == 1)
atomic_read       771 mm/oom_kill.c  			!atomic_read(&oom_victims), timeout);
atomic_read       833 mm/oom_kill.c  	if (atomic_read(&mm->mm_users) <= 1)
atomic_read      1014 mm/page_alloc.c 	if (unlikely(atomic_read(&page->_mapcount) != -1))
atomic_read      1036 mm/page_alloc.c 	if (unlikely(atomic_read(&page->_mapcount) != -1))
atomic_read      2048 mm/page_alloc.c 	if (unlikely(atomic_read(&page->_mapcount) != -1))
atomic_read       100 mm/rmap.c      	VM_BUG_ON(atomic_read(&anon_vma->refcount));
atomic_read      1104 mm/shmem.c     				       !atomic_read(&info->stop_eviction));
atomic_read      1107 mm/shmem.c     			if (!atomic_read(&info->stop_eviction))
atomic_read      4094 mm/slab.c      		unsigned long allochit = atomic_read(&cachep->allochit);
atomic_read      4095 mm/slab.c      		unsigned long allocmiss = atomic_read(&cachep->allocmiss);
atomic_read      4096 mm/slab.c      		unsigned long freehit = atomic_read(&cachep->freehit);
atomic_read      4097 mm/slab.c      		unsigned long freemiss = atomic_read(&cachep->freemiss);
atomic_read       301 mm/swap_state.c 	return READ_ONCE(enable_vma_readahead) && !atomic_read(&nr_rotate_swap);
atomic_read       513 mm/swap_state.c 				  atomic_read(&last_readahead_pages));
atomic_read      1619 mm/swapfile.c  		mapcount = atomic_read(&page[i]._mapcount) + 1;
atomic_read      2700 mm/swapfile.c  	if (seq->poll_event != atomic_read(&proc_poll_event)) {
atomic_read      2701 mm/swapfile.c  		seq->poll_event = atomic_read(&proc_poll_event);
atomic_read      2795 mm/swapfile.c  	seq->poll_event = atomic_read(&proc_poll_event);
atomic_read       641 mm/util.c      		return atomic_read(&page->_mapcount) >= 0;
atomic_read       643 mm/util.c      	if (atomic_read(compound_mapcount_ptr(page)) >= 0)
atomic_read       648 mm/util.c      		if (atomic_read(&page[i]._mapcount) >= 0)
atomic_read       706 mm/util.c      	ret = atomic_read(&page->_mapcount) + 1;
atomic_read       714 mm/util.c      	ret += atomic_read(compound_mapcount_ptr(page)) + 1;
atomic_read        64 mm/zpool.c     	refcount = atomic_read(&driver->refcount);
atomic_read       430 net/9p/trans_rdma.c 	if (unlikely(atomic_read(&rdma->excess_rc) > 0)) {
atomic_read       323 net/9p/trans_virtio.c 		if (atomic_read(&vp_pinned) >= chan->p9_max_pages) {
atomic_read       325 net/9p/trans_virtio.c 			      (atomic_read(&vp_pinned) < chan->p9_max_pages));
atomic_read        18 net/atm/atm_misc.c 	if (atomic_read(&sk_atm(vcc)->sk_rmem_alloc) <= sk_atm(vcc)->sk_rcvbuf)
atomic_read        33 net/atm/atm_misc.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) {
atomic_read        90 net/atm/atm_misc.c #define __HANDLE_ITEM(i) to->i = atomic_read(&from->i)
atomic_read       264 net/atm/br2684.c 		if (unlikely(atomic_read(&brvcc->qspace) > 0))
atomic_read       278 net/atm/br2684.c 	if (atomic_read(&brvcc->qspace) > 0)
atomic_read        80 net/atm/common.c 	if (atomic_read(&sk->sk_rmem_alloc))
atomic_read        82 net/atm/common.c 		       __func__, atomic_read(&sk->sk_rmem_alloc));
atomic_read       559 net/atm/common.c 		pr_debug("%d -= %d\n", atomic_read(&sk->sk_rmem_alloc),
atomic_read        48 net/atm/proc.c 		   atomic_read(&stats->tx), atomic_read(&stats->tx_err),
atomic_read        49 net/atm/proc.c 		   atomic_read(&stats->rx), atomic_read(&stats->rx_err),
atomic_read        50 net/atm/proc.c 		   atomic_read(&stats->rx_drop));
atomic_read       164 net/atm/resources.c #define __HANDLE_ITEM(i) to->i = atomic_read(&from->i)
atomic_read       250 net/ax25/af_ax25.c 		    atomic_read(&s->sk->sk_rmem_alloc) <= s->sk->sk_rcvbuf) {
atomic_read       124 net/ax25/ax25_ds_timer.c 			if (atomic_read(&sk->sk_rmem_alloc) <
atomic_read       263 net/ax25/ax25_in.c 				if (atomic_read(&sk->sk_rmem_alloc) >=
atomic_read        62 net/ax25/ax25_std_timer.c 			if (atomic_read(&sk->sk_rmem_alloc) <
atomic_read       282 net/batman-adv/bat_iv_ogm.c 	msecs = atomic_read(&bat_priv->orig_interval) - BATADV_JITTER;
atomic_read       297 net/batman-adv/bat_iv_ogm.c 	int hop_penalty = atomic_read(&bat_priv->hop_penalty);
atomic_read       550 net/batman-adv/bat_iv_ogm.c 	if (atomic_read(&bat_priv->aggregated_ogms) &&
atomic_read       641 net/batman-adv/bat_iv_ogm.c 	if (atomic_read(&bat_priv->aggregated_ogms) && !own_packet) {
atomic_read       667 net/batman-adv/bat_iv_ogm.c 		if (!own_packet && atomic_read(&bat_priv->aggregated_ogms))
atomic_read       821 net/batman-adv/bat_iv_ogm.c 	seqno = (u32)atomic_read(&hard_iface->bat_iv.ogm_seqno);
atomic_read      1592 net/batman-adv/bat_iv_ogm.c 	if_incoming_seqno = atomic_read(&if_incoming->bat_iv.ogm_seqno);
atomic_read      1705 net/batman-adv/bat_iv_ogm.c 	if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING) {
atomic_read      2438 net/batman-adv/bat_iv_ogm.c 		switch (atomic_read(&bat_priv->gw.sel_class)) {
atomic_read      2501 net/batman-adv/bat_iv_ogm.c 	if (atomic_read(&bat_priv->gw.sel_class) <= 2)
atomic_read      2536 net/batman-adv/bat_iv_ogm.c 	if ((atomic_read(&bat_priv->gw.sel_class) > 3) &&
atomic_read      2537 net/batman-adv/bat_iv_ogm.c 	    (orig_tq_avg - gw_tq_avg < atomic_read(&bat_priv->gw.sel_class)))
atomic_read       679 net/batman-adv/bat_v.c 	old_class = atomic_read(&bat_priv->gw.sel_class);
atomic_read       690 net/batman-adv/bat_v.c 	u32 class = atomic_read(&bat_priv->gw.sel_class);
atomic_read       792 net/batman-adv/bat_v.c 	threshold = atomic_read(&bat_priv->gw.sel_class);
atomic_read        51 net/batman-adv/bat_v_elp.c 	msecs = atomic_read(&hard_iface->bat_v.elp_interval) - BATADV_JITTER;
atomic_read        77 net/batman-adv/bat_v_elp.c 	throughput =  atomic_read(&hard_iface->bat_v.throughput_override);
atomic_read       270 net/batman-adv/bat_v_elp.c 	if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING)
atomic_read       287 net/batman-adv/bat_v_elp.c 	elp_packet->seqno = htonl(atomic_read(&hard_iface->bat_v.elp_seqno));
atomic_read       288 net/batman-adv/bat_v_elp.c 	elp_interval = atomic_read(&hard_iface->bat_v.elp_interval);
atomic_read       294 net/batman-adv/bat_v_elp.c 		   atomic_read(&hard_iface->bat_v.elp_seqno));
atomic_read       109 net/batman-adv/bat_v_ogm.c 	msecs = atomic_read(&bat_priv->orig_interval) - BATADV_JITTER;
atomic_read       245 net/batman-adv/bat_v_ogm.c 	if (!atomic_read(&bat_priv->aggregated_ogms)) {
atomic_read       275 net/batman-adv/bat_v_ogm.c 	if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING)
atomic_read       299 net/batman-adv/bat_v_ogm.c 	ogm_packet->seqno = htonl(atomic_read(&bat_priv->bat_v.ogm_seqno));
atomic_read       479 net/batman-adv/bat_v_ogm.c 	int hop_penalty = atomic_read(&bat_priv->hop_penalty);
atomic_read       646 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&backbone_gw->request_sent)) {
atomic_read       869 net/batman-adv/bridge_loop_avoidance.c 		if (atomic_read(&backbone_gw->request_sent)) {
atomic_read      1245 net/batman-adv/bridge_loop_avoidance.c 			if (atomic_read(&backbone_gw->request_sent))
atomic_read      1337 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&bat_priv->bridge_loop_avoidance))
atomic_read      1441 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&bat_priv->bridge_loop_avoidance))
atomic_read      1489 net/batman-adv/bridge_loop_avoidance.c 			if (atomic_read(&backbone_gw->request_sent) == 0)
atomic_read      1668 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&bat_priv->bridge_loop_avoidance))
atomic_read      1706 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&orig_node->bat_priv->bridge_loop_avoidance))
atomic_read      1839 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&bat_priv->bridge_loop_avoidance))
atomic_read      1845 net/batman-adv/bridge_loop_avoidance.c 	if (unlikely(atomic_read(&bat_priv->bla.num_requests)))
atomic_read      1950 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&bat_priv->bridge_loop_avoidance))
atomic_read      1958 net/batman-adv/bridge_loop_avoidance.c 	if (unlikely(atomic_read(&bat_priv->bla.num_requests)))
atomic_read      2177 net/batman-adv/bridge_loop_avoidance.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
atomic_read      2415 net/batman-adv/bridge_loop_avoidance.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
atomic_read      2518 net/batman-adv/bridge_loop_avoidance.c 	if (!atomic_read(&bat_priv->bridge_loop_avoidance))
atomic_read       744 net/batman-adv/distributed-arp-table.c 	dat_mode = atomic_read(&bat_priv->distributed_arp_table);
atomic_read       961 net/batman-adv/distributed-arp-table.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
atomic_read      1204 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read      1300 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read      1363 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read      1411 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read      1782 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read      1812 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read      1857 net/batman-adv/distributed-arp-table.c 	if (!atomic_read(&bat_priv->distributed_arp_table))
atomic_read       186 net/batman-adv/gateway_client.c 	if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT)
atomic_read       218 net/batman-adv/gateway_client.c 	if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT)
atomic_read       768 net/batman-adv/gateway_client.c 	switch (atomic_read(&bat_priv->gw.mode)) {
atomic_read       140 net/batman-adv/gateway_common.c 	gw_mode = atomic_read(&bat_priv->gw.mode);
atomic_read       148 net/batman-adv/gateway_common.c 		down = atomic_read(&bat_priv->gw.bandwidth_down);
atomic_read       149 net/batman-adv/gateway_common.c 		up = atomic_read(&bat_priv->gw.bandwidth_up);
atomic_read       177 net/batman-adv/gateway_common.c 	down_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_down);
atomic_read       178 net/batman-adv/gateway_common.c 	up_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_up);
atomic_read       246 net/batman-adv/gateway_common.c 	    atomic_read(&bat_priv->gw.mode) == BATADV_GW_MODE_CLIENT)
atomic_read       587 net/batman-adv/hard-interface.c 	if (atomic_read(&bat_priv->fragmentation) == 0)
atomic_read       777 net/batman-adv/hard-interface.c 	if (atomic_read(&bat_priv->fragmentation) &&
atomic_read       784 net/batman-adv/hard-interface.c 	if (!atomic_read(&bat_priv->fragmentation) &&
atomic_read       230 net/batman-adv/icmp_socket.c 		if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read        82 net/batman-adv/log.h 		if (atomic_read(&__batpriv->log_level) & (type) &&	\
atomic_read       485 net/batman-adv/main.c 	if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read       682 net/batman-adv/main.c 		ap_isolation_enabled = atomic_read(&vlan->ap_isolation);
atomic_read      1104 net/batman-adv/multicast.c 	if (!atomic_read(&bat_priv->multicast_mode))
atomic_read      1139 net/batman-adv/multicast.c 		return atomic_read(&bat_priv->mcast.num_want_all_ipv4);
atomic_read      1141 net/batman-adv/multicast.c 		return atomic_read(&bat_priv->mcast.num_want_all_ipv6);
atomic_read      1163 net/batman-adv/multicast.c 		return atomic_read(&bat_priv->mcast.num_want_all_rtr4);
atomic_read      1165 net/batman-adv/multicast.c 		return atomic_read(&bat_priv->mcast.num_want_all_rtr6);
atomic_read      1404 net/batman-adv/multicast.c 			atomic_read(&bat_priv->mcast.num_want_all_unsnoopables);
atomic_read      1428 net/batman-adv/multicast.c 		mcast_fanout = atomic_read(&bat_priv->multicast_fanout);
atomic_read      2259 net/batman-adv/multicast.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
atomic_read       187 net/batman-adv/netlink.c 	ap_isolation = atomic_read(&vlan->ap_isolation);
atomic_read       249 net/batman-adv/netlink.c 		       (u8)atomic_read(&bat_priv->tt.vn)))
atomic_read       275 net/batman-adv/netlink.c 		       !!atomic_read(&bat_priv->aggregated_ogms)))
atomic_read       290 net/batman-adv/netlink.c 		       !!atomic_read(&bat_priv->bonding)))
atomic_read       295 net/batman-adv/netlink.c 		       !!atomic_read(&bat_priv->bridge_loop_avoidance)))
atomic_read       301 net/batman-adv/netlink.c 		       !!atomic_read(&bat_priv->distributed_arp_table)))
atomic_read       306 net/batman-adv/netlink.c 		       !!atomic_read(&bat_priv->fragmentation)))
atomic_read       310 net/batman-adv/netlink.c 			atomic_read(&bat_priv->gw.bandwidth_down)))
atomic_read       314 net/batman-adv/netlink.c 			atomic_read(&bat_priv->gw.bandwidth_up)))
atomic_read       318 net/batman-adv/netlink.c 		       atomic_read(&bat_priv->gw.mode)))
atomic_read       327 net/batman-adv/netlink.c 				atomic_read(&bat_priv->gw.sel_class)))
atomic_read       332 net/batman-adv/netlink.c 		       atomic_read(&bat_priv->hop_penalty)))
atomic_read       337 net/batman-adv/netlink.c 			atomic_read(&bat_priv->log_level)))
atomic_read       343 net/batman-adv/netlink.c 		       !atomic_read(&bat_priv->multicast_mode)))
atomic_read       347 net/batman-adv/netlink.c 			atomic_read(&bat_priv->multicast_fanout)))
atomic_read       353 net/batman-adv/netlink.c 		       !!atomic_read(&bat_priv->network_coding)))
atomic_read       358 net/batman-adv/netlink.c 			atomic_read(&bat_priv->orig_interval)))
atomic_read       831 net/batman-adv/netlink.c 			atomic_read(&hard_iface->bat_v.elp_interval)))
atomic_read       835 net/batman-adv/netlink.c 			atomic_read(&hard_iface->bat_v.throughput_override)))
atomic_read      1041 net/batman-adv/netlink.c 		       !!atomic_read(&vlan->ap_isolation)))
atomic_read        92 net/batman-adv/network-coding.c 	nc_mode = atomic_read(&bat_priv->network_coding);
atomic_read       277 net/batman-adv/network-coding.c 	if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read       293 net/batman-adv/network-coding.c 	if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read       314 net/batman-adv/network-coding.c 	if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read       604 net/batman-adv/network-coding.c 	if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_ACTIVE &&
atomic_read       642 net/batman-adv/network-coding.c 	if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_ACTIVE &&
atomic_read       907 net/batman-adv/network-coding.c 	if (!atomic_read(&bat_priv->network_coding))
atomic_read      1512 net/batman-adv/network-coding.c 	if (!atomic_read(&bat_priv->network_coding))
atomic_read      1565 net/batman-adv/network-coding.c 	if (!atomic_read(&bat_priv->network_coding))
atomic_read      1809 net/batman-adv/network-coding.c 	if (!atomic_read(&bat_priv->network_coding))
atomic_read       565 net/batman-adv/routing.c 	if (!(recv_if == BATADV_IF_DEFAULT && atomic_read(&bat_priv->bonding)))
atomic_read       776 net/batman-adv/routing.c 		orig_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
atomic_read       787 net/batman-adv/routing.c 		orig_ttvn = (u8)atomic_read(&orig_node->last_ttvn);
atomic_read       854 net/batman-adv/routing.c 	curr_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
atomic_read       865 net/batman-adv/routing.c 		curr_ttvn = (u8)atomic_read(&orig_node->last_ttvn);
atomic_read       195 net/batman-adv/send.c 	if (atomic_read(&bat_priv->fragmentation) &&
atomic_read       239 net/batman-adv/send.c 	u8 ttvn = (u8)atomic_read(&orig_node->last_ttvn);
atomic_read       871 net/batman-adv/send.c 	if (atomic_read(&bat_priv->mesh_state) == BATADV_MESH_DEACTIVATING) {
atomic_read       142 net/batman-adv/soft-interface.c 	if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read       206 net/batman-adv/soft-interface.c 	if (atomic_read(&bat_priv->mesh_state) != BATADV_MESH_ACTIVE)
atomic_read       271 net/batman-adv/soft-interface.c 	gw_mode = atomic_read(&bat_priv->gw.mode);
atomic_read       162 net/batman-adv/sysfs.c 		       atomic_read(&bat_priv->_name) == 0 ?		\
atomic_read       202 net/batman-adv/sysfs.c 	return sprintf(buff, "%i\n", atomic_read(&bat_priv->_var));	\
atomic_read       244 net/batman-adv/sysfs.c 			     atomic_read(&vlan->_name) == 0 ?		\
atomic_read       302 net/batman-adv/sysfs.c 	length = sprintf(buff, "%i\n", atomic_read(&hard_iface->_var));	\
atomic_read       346 net/batman-adv/sysfs.c 	if (atomic_read(attr) == enabled)
atomic_read       350 net/batman-adv/sysfs.c 		    atomic_read(attr) == 1 ? "enabled" : "disabled",
atomic_read       406 net/batman-adv/sysfs.c 	if (atomic_read(attr) == uint_val)
atomic_read       413 net/batman-adv/sysfs.c 		    attr_name, ifname, atomic_read(attr), uint_val);
atomic_read       468 net/batman-adv/sysfs.c 	switch (atomic_read(&bat_priv->gw.mode)) {
atomic_read       526 net/batman-adv/sysfs.c 	if (atomic_read(&bat_priv->gw.mode) == gw_mode_tmp)
atomic_read       529 net/batman-adv/sysfs.c 	switch (atomic_read(&bat_priv->gw.mode)) {
atomic_read       583 net/batman-adv/sysfs.c 	return sprintf(buff, "%i\n", atomic_read(&bat_priv->gw.sel_class));
atomic_read       627 net/batman-adv/sysfs.c 	down = atomic_read(&bat_priv->gw.bandwidth_down);
atomic_read       628 net/batman-adv/sysfs.c 	up = atomic_read(&bat_priv->gw.bandwidth_up);
atomic_read      1155 net/batman-adv/sysfs.c 	old_tp_override = atomic_read(&hard_iface->bat_v.throughput_override);
atomic_read      1191 net/batman-adv/sysfs.c 	tp_override = atomic_read(&hard_iface->bat_v.throughput_override);
atomic_read       467 net/batman-adv/tp_meter.c 	if (unlikely(atomic_read(&tp_vars->sending) == 0))
atomic_read       487 net/batman-adv/tp_meter.c 	if (atomic_read(&tp_vars->sending) == 0)
atomic_read       511 net/batman-adv/tp_meter.c 		   atomic_read(&tp_vars->last_acked));
atomic_read       518 net/batman-adv/tp_meter.c 	tp_vars->last_sent = atomic_read(&tp_vars->last_acked);
atomic_read       646 net/batman-adv/tp_meter.c 	if (unlikely(atomic_read(&tp_vars->sending) == 0))
atomic_read       651 net/batman-adv/tp_meter.c 			      (u32)atomic_read(&tp_vars->last_acked)))
atomic_read       673 net/batman-adv/tp_meter.c 	if (atomic_read(&tp_vars->last_acked) == recv_ack) {
atomic_read       675 net/batman-adv/tp_meter.c 		if (atomic_read(&tp_vars->dup_acks) != 3)
atomic_read       708 net/batman-adv/tp_meter.c 		atomic64_add(recv_ack - atomic_read(&tp_vars->last_acked),
atomic_read       741 net/batman-adv/tp_meter.c 		if (recv_ack - atomic_read(&tp_vars->last_acked) >= mss)
atomic_read       770 net/batman-adv/tp_meter.c 	win_limit = atomic_read(&tp_vars->last_acked) + tp_vars->cwnd;
atomic_read       848 net/batman-adv/tp_meter.c 	while (atomic_read(&tp_vars->sending) != 0) {
atomic_read       316 net/batman-adv/translation-table.c 	count = atomic_read(&tt_global_entry->orig_list_count);
atomic_read       573 net/batman-adv/translation-table.c 		tt_local_entries += atomic_read(&vlan->tt.num_entries);
atomic_read       703 net/batman-adv/translation-table.c 	packet_size_max = atomic_read(&bat_priv->packet_size_max);
atomic_read       729 net/batman-adv/translation-table.c 		   (u8)atomic_read(&bat_priv->tt.vn));
atomic_read       867 net/batman-adv/translation-table.c 		num_entries += atomic_read(&vlan->tt.num_entries);
atomic_read       887 net/batman-adv/translation-table.c 	(*tt_data)->ttvn = atomic_read(&orig_node->last_ttvn);
atomic_read       941 net/batman-adv/translation-table.c 		vlan_entries = atomic_read(&vlan->tt.num_entries);
atomic_read       966 net/batman-adv/translation-table.c 	(*tt_data)->ttvn = atomic_read(&bat_priv->tt.vn);
atomic_read       971 net/batman-adv/translation-table.c 		vlan_entries = atomic_read(&vlan->tt.num_entries);
atomic_read      1004 net/batman-adv/translation-table.c 	tt_diff_entries_num = atomic_read(&bat_priv->tt.local_changes);
atomic_read      1096 net/batman-adv/translation-table.c 		   net_dev->name, (u8)atomic_read(&bat_priv->tt.vn));
atomic_read      1225 net/batman-adv/translation-table.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
atomic_read      1919 net/batman-adv/translation-table.c 		last_ttvn = atomic_read(&best_entry->orig_node->last_ttvn);
atomic_read      1951 net/batman-adv/translation-table.c 		last_ttvn = atomic_read(&orig_entry->orig_node->last_ttvn);
atomic_read      2053 net/batman-adv/translation-table.c 	last_ttvn = atomic_read(&orig->orig_node->last_ttvn);
atomic_read      3257 net/batman-adv/translation-table.c 	orig_ttvn = (u8)atomic_read(&req_dst_orig_node->last_ttvn);
atomic_read      3313 net/batman-adv/translation-table.c 	if (tt_len > atomic_read(&bat_priv->packet_size_max)) {
atomic_read      3381 net/batman-adv/translation-table.c 	my_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
atomic_read      3420 net/batman-adv/translation-table.c 		req_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
atomic_read      3929 net/batman-adv/translation-table.c 	if (atomic_read(&bat_priv->tt.local_changes) < 1) {
atomic_read      3944 net/batman-adv/translation-table.c 		   (u8)atomic_read(&bat_priv->tt.vn));
atomic_read      3984 net/batman-adv/translation-table.c 	if (!atomic_read(&vlan->ap_isolation))
atomic_read      4023 net/batman-adv/translation-table.c 	u8 orig_ttvn = (u8)atomic_read(&orig_node->last_ttvn);
atomic_read      4165 net/batman-adv/translation-table.c 				  atomic_read(&orig_node->last_ttvn)))
atomic_read      4186 net/batman-adv/translation-table.c 	int packet_size_max = atomic_read(&bat_priv->packet_size_max);
atomic_read      4380 net/batman-adv/translation-table.c 			     atomic_read(&orig_node->last_ttvn) + 1);
atomic_read       110 net/bluetooth/6lowpan.c 	BT_DBG("peers %d addr %pMR type %d", atomic_read(&dev->peer_count),
atomic_read       164 net/bluetooth/6lowpan.c 	int count = atomic_read(&dev->peer_count);
atomic_read       821 net/bluetooth/6lowpan.c 	if (!err && last && dev && !atomic_read(&dev->peer_count)) {
atomic_read       492 net/bluetooth/bnep/core.c 		if (atomic_read(&s->terminate))
atomic_read       291 net/bluetooth/cmtp/core.c 		if (atomic_read(&session->terminate))
atomic_read       400 net/bluetooth/hci_conn.c 	int refcnt = atomic_read(&conn->refcnt);
atomic_read      3594 net/bluetooth/hci_core.c 	if (atomic_read(&hdev->promisc)) {
atomic_read      4460 net/bluetooth/hci_core.c 		if (atomic_read(&hdev->promisc)) {
atomic_read      4517 net/bluetooth/hci_core.c 	       atomic_read(&hdev->cmd_cnt), skb_queue_len(&hdev->cmd_q));
atomic_read      4520 net/bluetooth/hci_core.c 	if (atomic_read(&hdev->cmd_cnt)) {
atomic_read      3485 net/bluetooth/hci_event.c 	if (atomic_read(&hdev->cmd_cnt) && !skb_queue_empty(&hdev->cmd_q))
atomic_read      3598 net/bluetooth/hci_event.c 	if (atomic_read(&hdev->cmd_cnt) && !skb_queue_empty(&hdev->cmd_q))
atomic_read       303 net/bluetooth/hci_sock.c 	if (!atomic_read(&monitor_promisc))
atomic_read       731 net/bluetooth/hci_sock.c 	if (atomic_read(&monitor_promisc)) {
atomic_read       108 net/bluetooth/hidp/core.c 	if (atomic_read(&session->terminate))
atomic_read       244 net/bluetooth/hidp/core.c 	if (atomic_read(&session->terminate))
atomic_read       276 net/bluetooth/hidp/core.c 	       !atomic_read(&session->terminate)) {
atomic_read       281 net/bluetooth/hidp/core.c 				|| atomic_read(&session->terminate),
atomic_read       351 net/bluetooth/hidp/core.c 	       !atomic_read(&session->terminate)) {
atomic_read       356 net/bluetooth/hidp/core.c 				|| atomic_read(&session->terminate),
atomic_read      1065 net/bluetooth/hidp/core.c 	while (atomic_read(&session->state) <= HIDP_SESSION_IDLING)
atomic_read      1067 net/bluetooth/hidp/core.c 			   atomic_read(&session->state) > HIDP_SESSION_IDLING);
atomic_read      1172 net/bluetooth/hidp/core.c 	    atomic_read(&session->state) > HIDP_SESSION_PREPARING)
atomic_read      1205 net/bluetooth/hidp/core.c 		if (atomic_read(&session->terminate))
atomic_read      1036 net/bluetooth/l2cap_sock.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf >> 1)
atomic_read      1331 net/bluetooth/l2cap_sock.c 	lock_sock_nested(sk, atomic_read(&chan->nesting));
atomic_read        60 net/bluetooth/rfcomm/sock.c 	if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf)
atomic_read       643 net/bluetooth/rfcomm/sock.c 	if (atomic_read(&sk->sk_rmem_alloc) <= (sk->sk_rcvbuf >> 2))
atomic_read       356 net/bluetooth/rfcomm/tty.c 	int pending = 40 - atomic_read(&dev->wmem_alloc);
atomic_read       449 net/bridge/br_multicast.c 	if (atomic_read(&br->mdb_hash_tbl.nelems) >= br->hash_max) {
atomic_read       132 net/caif/caif_socket.c 	if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >=
atomic_read       135 net/caif/caif_socket.c 				    atomic_read(&cf_sk->sk.sk_rmem_alloc),
atomic_read       261 net/caif/caif_socket.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk_rcvbuf_lowwater(cf_sk)) {
atomic_read       134 net/caif/cfctrl.c 	req->sequence_no = atomic_read(&ctrl->req_seq_no);
atomic_read        65 net/can/j1939/socket.c 	return atomic_read(&jsk->skb_pending);
atomic_read       345 net/ceph/debugfs.c 		   atomic_read(&osdc->num_requests),
atomic_read       346 net/ceph/debugfs.c 		   atomic_read(&osdc->num_homeless));
atomic_read       370 net/ceph/messenger.c 	if (atomic_read(&con->msgr->stopping)) {
atomic_read      3343 net/ceph/osd_client.c 	if (atomic_read(&osdc->num_homeless) || !list_empty(&slow_osds))
atomic_read      5224 net/ceph/osd_client.c 	WARN_ON(atomic_read(&osdc->num_requests));
atomic_read      5225 net/ceph/osd_client.c 	WARN_ON(atomic_read(&osdc->num_homeless));
atomic_read       103 net/core/bpf_sk_storage.c 	    atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
atomic_read      1775 net/core/dev.c 		wanted = atomic_read(&netstamp_wanted);
atomic_read      1795 net/core/dev.c 		wanted = atomic_read(&netstamp_wanted);
atomic_read      1199 net/core/filter.c 	    atomic_read(&sk->sk_omem_alloc) + filter_size < sysctl_optmem_max) {
atomic_read       223 net/core/neighbour.c 	int max_clean = atomic_read(&tbl->gc_entries) - tbl->gc_thresh2;
atomic_read       621 net/core/neighbour.c 	if (atomic_read(&tbl->entries) > (1 << nht->hash_shift))
atomic_read       911 net/core/neighbour.c 	if (atomic_read(&tbl->entries) < tbl->gc_thresh1)
atomic_read      1076 net/core/neighbour.c 	    atomic_read(&neigh->probes) >= neigh_max_probes(neigh)) {
atomic_read      1733 net/core/neighbour.c 	if (atomic_read(&tbl->entries))
atomic_read      2058 net/core/neighbour.c 			.ndtc_entries		= atomic_read(&tbl->entries),
atomic_read      2463 net/core/neighbour.c 	if (nla_put_u32(skb, NDA_PROBES, atomic_read(&neigh->probes)) ||
atomic_read      3313 net/core/neighbour.c 		   atomic_read(&tbl->entries),
atomic_read       294 net/core/net-sysfs.c 		       atomic_read(&netdev->carrier_up_count) +
atomic_read       295 net/core/net-sysfs.c 		       atomic_read(&netdev->carrier_down_count));
atomic_read       305 net/core/net-sysfs.c 	return sprintf(buf, fmt_dec, atomic_read(&netdev->carrier_up_count));
atomic_read       315 net/core/net-sysfs.c 	return sprintf(buf, fmt_dec, atomic_read(&netdev->carrier_down_count));
atomic_read       197 net/core/page_pool.c 	u32 release_cnt = atomic_read(&pool->pages_state_release_cnt);
atomic_read       812 net/core/rtnetlink.c 		ci.rta_clntref = atomic_read(&dst->__refcnt);
atomic_read      1635 net/core/rtnetlink.c 			atomic_read(&dev->carrier_up_count) +
atomic_read      1636 net/core/rtnetlink.c 			atomic_read(&dev->carrier_down_count)) ||
atomic_read      1639 net/core/rtnetlink.c 			atomic_read(&dev->carrier_up_count)) ||
atomic_read      1641 net/core/rtnetlink.c 			atomic_read(&dev->carrier_down_count)))
atomic_read      1155 net/core/skbuff.c 		next = (u32)atomic_read(&sk->sk_zckey);
atomic_read      4417 net/core/skbuff.c 	if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >=
atomic_read       703 net/core/skmsg.c 		if (atomic_read(&sk_other->sk_rmem_alloc) <=
atomic_read       724 net/core/skmsg.c 		     atomic_read(&sk_other->sk_rmem_alloc) <=
atomic_read       456 net/core/sock.c 	if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf) {
atomic_read      1712 net/core/sock.c 	if (atomic_read(&sk->sk_omem_alloc))
atomic_read      1714 net/core/sock.c 			 __func__, atomic_read(&sk->sk_omem_alloc));
atomic_read      2122 net/core/sock.c 	if (atomic_read(&sk->sk_omem_alloc) + SKB_TRUESIZE(size) >
atomic_read      2142 net/core/sock.c 	    atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
atomic_read      2526 net/core/sock.c 		if (atomic_read(&sk->sk_rmem_alloc) < sk_get_rmem0(sk, prot))
atomic_read      2548 net/core/sock.c 				 atomic_read(&sk->sk_rmem_alloc) +
atomic_read      3219 net/core/sock.c 	mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc);
atomic_read      3221 net/core/sock.c 	mem[SK_MEMINFO_DROPS] = atomic_read(&sk->sk_drops);
atomic_read       344 net/dccp/proto.c 		if (atomic_read(&sk->sk_rmem_alloc) > 0)
atomic_read       585 net/decnet/dn_nsp_in.c 	if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >=
atomic_read       193 net/decnet/dn_route.c 			if (atomic_read(&rt->dst.__refcnt) > 1 ||
atomic_read       227 net/decnet/dn_route.c 			if (atomic_read(&rt->dst.__refcnt) > 1 ||
atomic_read      1838 net/decnet/dn_route.c 		   atomic_read(&rt->dst.__refcnt),
atomic_read       153 net/ipv4/af_inet.c 	WARN_ON(atomic_read(&sk->sk_rmem_alloc));
atomic_read       336 net/ipv4/arp.c 	int probes = atomic_read(&neigh->probes);
atomic_read      1852 net/ipv4/devinet.c 		cb->seq = atomic_read(&tgt_net->ipv4.dev_addr_genid) ^
atomic_read      2252 net/ipv4/devinet.c 		cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
atomic_read      1290 net/ipv4/fib_semantics.c 	nh->nh_saddr_genid = atomic_read(&net->ipv4.dev_addr_genid);
atomic_read      1306 net/ipv4/fib_semantics.c 		if (nh->nh_saddr_genid == atomic_read(&net->ipv4.dev_addr_genid))
atomic_read      2180 net/ipv4/fib_semantics.c 		if (hash > atomic_read(&nexthop_nh->fib_nh_upper_bound))
atomic_read       505 net/ipv4/inet_connection_sock.c 				   atomic_read(&newsk->sk_rmem_alloc));
atomic_read       161 net/ipv4/inet_timewait_sock.c 	if (atomic_read(&dr->tw_count) >= dr->sysctl_max_tw_buckets)
atomic_read      1168 net/ipv4/ipmr.c 		if (atomic_read(&mrt->cache_resolve_queue_len) == 1)
atomic_read      1331 net/ipv4/ipmr.c 		if (atomic_read(&mrt->cache_resolve_queue_len) != 0) {
atomic_read      2756 net/ipv4/ipmr.c 	u32 queue_len = atomic_read(&mrt->cache_resolve_queue_len);
atomic_read       504 net/ipv4/nexthop.c 		if (hash > atomic_read(&nhge->upper_bound))
atomic_read      1119 net/ipv4/ping.c 		atomic_read(&sp->sk_drops));
atomic_read        61 net/ipv4/proc.c 		   atomic_read(&net->ipv4.tcp_death_row.tw_count), sockets,
atomic_read        71 net/ipv4/proc.c 		   atomic_read(&net->ipv4.fqdir->rhashtable.nelems),
atomic_read      1082 net/ipv4/raw.c 		refcount_read(&sp->sk_refcnt), sp, atomic_read(&sp->sk_drops));
atomic_read      1557 net/ipv4/tcp.c 		      !atomic_read(&sk->sk_rmem_alloc)))
atomic_read       496 net/ipv4/tcp_fastopen.c 	int tfo_da_times = atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times);
atomic_read       540 net/ipv4/tcp_fastopen.c 		   atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times)) {
atomic_read       488 net/ipv4/tcp_input.c 			   min(atomic_read(&sk->sk_rmem_alloc),
atomic_read       491 net/ipv4/tcp_input.c 	if (atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf)
atomic_read      4531 net/ipv4/tcp_input.c 	if (atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf ||
atomic_read      5093 net/ipv4/tcp_input.c 			if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf &&
atomic_read      5125 net/ipv4/tcp_input.c 	if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf)
atomic_read      5130 net/ipv4/tcp_input.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf)
atomic_read      5141 net/ipv4/tcp_input.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf)
atomic_read      5149 net/ipv4/tcp_input.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf)
atomic_read      1457 net/ipv4/udp.c 	rmem = atomic_read(&sk->sk_rmem_alloc);
atomic_read      2920 net/ipv4/udp.c 		atomic_read(&sp->sk_drops));
atomic_read       733 net/ipv6/addrconf.c 		cb->seq = atomic_read(&net->ipv6.dev_addr_genid) ^
atomic_read      5201 net/ipv6/addrconf.c 	cb->seq = atomic_read(&tgt_net->ipv6.dev_addr_genid) ^ tgt_net->dev_base_seq;
atomic_read      1049 net/ipv6/datagram.c 		   atomic_read(&sp->sk_drops));
atomic_read        95 net/ipv6/ip6_fib.c 		old = atomic_read(&net->ipv6.fib6_sernum);
atomic_read       163 net/ipv6/ip6_flowlabel.c 			if (atomic_read(&fl->users) == 0) {
atomic_read       180 net/ipv6/ip6_flowlabel.c 	if (!sched && atomic_read(&fl_size))
atomic_read       201 net/ipv6/ip6_flowlabel.c 			    atomic_read(&fl->users) == 0) {
atomic_read       469 net/ipv6/ip6_flowlabel.c 	int room = FL_MAX_SIZE - atomic_read(&fl_size);
atomic_read       822 net/ipv6/ip6_flowlabel.c 			   atomic_read(&fl->users),
atomic_read      1529 net/ipv6/ip6mr.c 		if (atomic_read(&mrt->cache_resolve_queue_len) != 0) {
atomic_read       721 net/ipv6/ndisc.c 	int probes = atomic_read(&neigh->probes);
atomic_read        47 net/ipv6/proc.c 		   atomic_read(&net->ipv6.fqdir->rhashtable.nelems),
atomic_read       449 net/ipv6/route.c 	if (fl6->mp_hash <= atomic_read(&match->fib6_nh->fib_nh_upper_bound))
atomic_read       457 net/ipv6/route.c 		nh_upper_bound = atomic_read(&nh->fib_nh_upper_bound);
atomic_read      6048 net/ipv6/route.c 		   atomic_read(&net->ipv6.rt6_stats->fib_rt_alloc),
atomic_read       295 net/iucv/af_iucv.c 		return ((atomic_read(&iucv->msg_sent) < iucv->msglimit_peer) &&
atomic_read       296 net/iucv/af_iucv.c 			(atomic_read(&iucv->pendings) <= 0));
atomic_read       335 net/iucv/af_iucv.c 		confirm_recv = atomic_read(&iucv->msg_recv);
atomic_read       386 net/iucv/af_iucv.c 		WARN_ON(atomic_read(&iucv->msg_recv) < 0);
atomic_read       418 net/iucv/af_iucv.c 	WARN_ON(atomic_read(&sk->sk_rmem_alloc));
atomic_read      1456 net/iucv/af_iucv.c 			if (atomic_read(&iucv->msg_recv) > iucv->msglimit) {
atomic_read      1479 net/iucv/af_iucv.c 			if (atomic_read(&iucv->msg_recv) >=
atomic_read      1860 net/iucv/af_iucv.c 	len = atomic_read(&sk->sk_rmem_alloc);
atomic_read      2312 net/iucv/af_iucv.c 				if (atomic_read(&iucv->pendings) <= 0)
atomic_read       150 net/kcm/kcmproc.c 		   atomic_read(&psock->sk->sk_rmem_alloc),
atomic_read       192 net/kcm/kcmsock.c 	if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf)
atomic_read        76 net/key/af_key.c 	if (3 * atomic_read(&sk->sk_rmem_alloc) <= 2 * sk->sk_rcvbuf)
atomic_read       107 net/key/af_key.c 	WARN_ON(atomic_read(&sk->sk_rmem_alloc));
atomic_read       200 net/key/af_key.c 	if (atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf)
atomic_read      3057 net/key/af_key.c 	if (atomic_read(&net_pfkey->socks_nr) == 0)
atomic_read      3716 net/key/af_key.c 	    3 * atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf)
atomic_read       926 net/llc/llc_conn.c 		__func__, atomic_read(&llc_sock_nr));
atomic_read       978 net/llc/llc_conn.c 			atomic_read(&llc_sock_nr));
atomic_read       982 net/llc/llc_conn.c 			__func__, atomic_read(&llc_sock_nr));
atomic_read      2661 net/mac80211/cfg.c 	       smps_mode, atomic_read(&sdata->u.ap.num_mcast_sta));
atomic_read       103 net/mac80211/debugfs_netdev.c 	return scnprintf(buf, buflen, "%d\n", atomic_read(&sdata->field));\
atomic_read      1584 net/mac80211/ieee80211_i.h 		return atomic_read(&sdata->u.ap.num_mcast_sta);
atomic_read      1586 net/mac80211/ieee80211_i.h 		return atomic_read(&sdata->u.vlan.num_mcast_sta);
atomic_read        14 net/mac80211/led.c 	if (!atomic_read(&local->assoc_led_active))
atomic_read        24 net/mac80211/led.c 	if (!atomic_read(&local->radio_led_active))
atomic_read        18 net/mac80211/led.h 	if (!atomic_read(&local->rx_led_active))
atomic_read        29 net/mac80211/led.h 	if (!atomic_read(&local->tx_led_active))
atomic_read        78 net/mac80211/led.h 	if (ieee80211_is_data(fc) && atomic_read(&local->tpt_led_active))
atomic_read        87 net/mac80211/led.h 	if (ieee80211_is_data(fc) && atomic_read(&local->tpt_led_active))
atomic_read        42 net/mac80211/main.c 	if (atomic_read(&local->iff_allmultis))
atomic_read       279 net/mac80211/mesh.c 	neighbors = atomic_read(&ifmsh->estab_plinks);
atomic_read       338 net/mac80211/mesh.h 	       atomic_read(&sdata->u.mesh.estab_plinks);
atomic_read       440 net/mac80211/tx.c 	if (!atomic_read(&ps->num_sta_ps) && skb_queue_empty(&ps->bc_buf))
atomic_read      4219 net/mac80211/tx.c 	if (atomic_read(&ps->num_sta_ps) > 0)
atomic_read       289 net/mac80211/util.c 	    (ps && atomic_read(&ps->num_sta_ps)) || ac != vif->txq->ac)
atomic_read       614 net/netfilter/ipset/ip_set_hash_gen.h 	if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
atomic_read      1008 net/netfilter/ipset/ip_set_hash_gen.h 	if (atomic_read(&t->ref) && ext->target) {
atomic_read      1035 net/netfilter/ipset/ip_set_hash_gen.h 	if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
atomic_read      1096 net/netfilter/ipset/ip_set_hash_gen.h 		if (atomic_read(&t->ref) && ext->target) {
atomic_read      1148 net/netfilter/ipset/ip_set_hash_gen.h 	if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
atomic_read      1336 net/netfilter/ipset/ip_set_hash_gen.h 		if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
atomic_read       585 net/netfilter/ipvs/ip_vs_app.c 			   atomic_read(&inc->usecnt),
atomic_read       298 net/netfilter/ipvs/ip_vs_conn.c 	if (!cp && atomic_read(&ip_vs_conn_no_cport_cnt)) {
atomic_read       563 net/netfilter/ipvs/ip_vs_conn.c 	return atomic_read(&dest->activeconns)
atomic_read       564 net/netfilter/ipvs/ip_vs_conn.c 		+ atomic_read(&dest->inactconns);
atomic_read       584 net/netfilter/ipvs/ip_vs_conn.c 	conn_flags = atomic_read(&dest->conn_flags);
atomic_read       681 net/netfilter/ipvs/ip_vs_conn.c 		if (pd && atomic_read(&pd->appcnt))
atomic_read       744 net/netfilter/ipvs/ip_vs_conn.c 		(atomic_read(&dest->weight) == 0);
atomic_read       818 net/netfilter/ipvs/ip_vs_conn.c 	if (atomic_read(&cp->n_control))
atomic_read       832 net/netfilter/ipvs/ip_vs_conn.c 			if (!cp->timeout && !atomic_read(&ct->n_control) &&
atomic_read       868 net/netfilter/ipvs/ip_vs_conn.c 		  atomic_read(&cp->n_control));
atomic_read       984 net/netfilter/ipvs/ip_vs_conn.c 	if (unlikely(pd && atomic_read(&pd->appcnt)))
atomic_read      1240 net/netfilter/ipvs/ip_vs_conn.c 	i = atomic_read(&cp->in_pkts);
atomic_read      1276 net/netfilter/ipvs/ip_vs_conn.c 			if (atomic_read(&cp->n_control))
atomic_read      1364 net/netfilter/ipvs/ip_vs_conn.c 	if (atomic_read(&ipvs->conn_count) != 0) {
atomic_read      1428 net/netfilter/ipvs/ip_vs_core.c 	if (atomic_read(&ipvs->conn_out_counter)) {
atomic_read      2067 net/netfilter/ipvs/ip_vs_core.c 		    unlikely(!atomic_read(&cp->dest->weight))) {
atomic_read      2072 net/netfilter/ipvs/ip_vs_core.c 			if (!atomic_read(&cp->n_control)) {
atomic_read      2084 net/netfilter/ipvs/ip_vs_core.c 			if (!atomic_read(&cp->n_control))
atomic_read      2363 net/netfilter/ipvs/ip_vs_core.c 	ipvs->gen = atomic_read(&ipvs_netns_cnt);
atomic_read       225 net/netfilter/ipvs/ip_vs_ctl.c 	if (atomic_read(&ipvs->dropentry))
atomic_read       425 net/netfilter/ipvs/ip_vs_ctl.c 	    atomic_read(&ipvs->ftpsvc_counter) &&
atomic_read       435 net/netfilter/ipvs/ip_vs_ctl.c 	    && atomic_read(&ipvs->nullsvc_counter)) {
atomic_read      2174 net/netfilter/ipvs/ip_vs_ctl.c 					   ip_vs_fwd_name(atomic_read(&dest->conn_flags)),
atomic_read      2175 net/netfilter/ipvs/ip_vs_ctl.c 					   atomic_read(&dest->weight),
atomic_read      2176 net/netfilter/ipvs/ip_vs_ctl.c 					   atomic_read(&dest->activeconns),
atomic_read      2177 net/netfilter/ipvs/ip_vs_ctl.c 					   atomic_read(&dest->inactconns));
atomic_read      2185 net/netfilter/ipvs/ip_vs_ctl.c 					   ip_vs_fwd_name(atomic_read(&dest->conn_flags)),
atomic_read      2186 net/netfilter/ipvs/ip_vs_ctl.c 					   atomic_read(&dest->weight),
atomic_read      2187 net/netfilter/ipvs/ip_vs_ctl.c 					   atomic_read(&dest->activeconns),
atomic_read      2188 net/netfilter/ipvs/ip_vs_ctl.c 					   atomic_read(&dest->inactconns));
atomic_read      2663 net/netfilter/ipvs/ip_vs_ctl.c 			entry.conn_flags = atomic_read(&dest->conn_flags);
atomic_read      2664 net/netfilter/ipvs/ip_vs_ctl.c 			entry.weight = atomic_read(&dest->weight);
atomic_read      2667 net/netfilter/ipvs/ip_vs_ctl.c 			entry.activeconns = atomic_read(&dest->activeconns);
atomic_read      2668 net/netfilter/ipvs/ip_vs_ctl.c 			entry.inactconns = atomic_read(&dest->inactconns);
atomic_read      2669 net/netfilter/ipvs/ip_vs_ctl.c 			entry.persistconns = atomic_read(&dest->persistconns);
atomic_read      3257 net/netfilter/ipvs/ip_vs_ctl.c 			(atomic_read(&dest->conn_flags) &
atomic_read      3260 net/netfilter/ipvs/ip_vs_ctl.c 			atomic_read(&dest->weight)) ||
atomic_read      3270 net/netfilter/ipvs/ip_vs_ctl.c 			atomic_read(&dest->activeconns)) ||
atomic_read      3272 net/netfilter/ipvs/ip_vs_ctl.c 			atomic_read(&dest->inactconns)) ||
atomic_read      3274 net/netfilter/ipvs/ip_vs_ctl.c 			atomic_read(&dest->persistconns)) ||
atomic_read       225 net/netfilter/ipvs/ip_vs_dh.c 	    || atomic_read(&dest->weight) <= 0
atomic_read        34 net/netfilter/ipvs/ip_vs_fo.c 		    atomic_read(&dest->weight) > hw) {
atomic_read        36 net/netfilter/ipvs/ip_vs_fo.c 			hw = atomic_read(&dest->weight);
atomic_read        44 net/netfilter/ipvs/ip_vs_fo.c 			      atomic_read(&hweight->activeconns),
atomic_read        45 net/netfilter/ipvs/ip_vs_fo.c 			      atomic_read(&hweight->weight));
atomic_read       311 net/netfilter/ipvs/ip_vs_lblc.c 	if (atomic_read(&tbl->entries) <= tbl->max_size) {
atomic_read       316 net/netfilter/ipvs/ip_vs_lblc.c 	goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3;
atomic_read       420 net/netfilter/ipvs/ip_vs_lblc.c 		if (atomic_read(&dest->weight) > 0) {
atomic_read       437 net/netfilter/ipvs/ip_vs_lblc.c 		if ((__s64)loh * atomic_read(&dest->weight) >
atomic_read       438 net/netfilter/ipvs/ip_vs_lblc.c 		    (__s64)doh * atomic_read(&least->weight)) {
atomic_read       448 net/netfilter/ipvs/ip_vs_lblc.c 		      atomic_read(&least->activeconns),
atomic_read       450 net/netfilter/ipvs/ip_vs_lblc.c 		      atomic_read(&least->weight), loh);
atomic_read       463 net/netfilter/ipvs/ip_vs_lblc.c 	if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) {
atomic_read       467 net/netfilter/ipvs/ip_vs_lblc.c 			if (atomic_read(&d->activeconns)*2
atomic_read       468 net/netfilter/ipvs/ip_vs_lblc.c 			    < atomic_read(&d->weight)) {
atomic_read       507 net/netfilter/ipvs/ip_vs_lblc.c 		    atomic_read(&dest->weight) > 0 && !is_overloaded(dest, svc))
atomic_read       173 net/netfilter/ipvs/ip_vs_lblcr.c 		if ((atomic_read(&least->weight) > 0)
atomic_read       189 net/netfilter/ipvs/ip_vs_lblcr.c 		if (((__s64)loh * atomic_read(&dest->weight) >
atomic_read       190 net/netfilter/ipvs/ip_vs_lblcr.c 		     (__s64)doh * atomic_read(&least->weight))
atomic_read       202 net/netfilter/ipvs/ip_vs_lblcr.c 		      atomic_read(&least->activeconns),
atomic_read       204 net/netfilter/ipvs/ip_vs_lblcr.c 		      atomic_read(&least->weight), loh);
atomic_read       222 net/netfilter/ipvs/ip_vs_lblcr.c 		if (atomic_read(&most->weight) > 0) {
atomic_read       235 net/netfilter/ipvs/ip_vs_lblcr.c 		if (((__s64)moh * atomic_read(&dest->weight) <
atomic_read       236 net/netfilter/ipvs/ip_vs_lblcr.c 		     (__s64)doh * atomic_read(&most->weight))
atomic_read       237 net/netfilter/ipvs/ip_vs_lblcr.c 		    && (atomic_read(&dest->weight) > 0)) {
atomic_read       247 net/netfilter/ipvs/ip_vs_lblcr.c 		      atomic_read(&most->activeconns),
atomic_read       249 net/netfilter/ipvs/ip_vs_lblcr.c 		      atomic_read(&most->weight), moh);
atomic_read       475 net/netfilter/ipvs/ip_vs_lblcr.c 	if (atomic_read(&tbl->entries) <= tbl->max_size) {
atomic_read       480 net/netfilter/ipvs/ip_vs_lblcr.c 	goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3;
atomic_read       584 net/netfilter/ipvs/ip_vs_lblcr.c 		if (atomic_read(&dest->weight) > 0) {
atomic_read       601 net/netfilter/ipvs/ip_vs_lblcr.c 		if ((__s64)loh * atomic_read(&dest->weight) >
atomic_read       602 net/netfilter/ipvs/ip_vs_lblcr.c 		    (__s64)doh * atomic_read(&least->weight)) {
atomic_read       612 net/netfilter/ipvs/ip_vs_lblcr.c 		      atomic_read(&least->activeconns),
atomic_read       614 net/netfilter/ipvs/ip_vs_lblcr.c 		      atomic_read(&least->weight), loh);
atomic_read       627 net/netfilter/ipvs/ip_vs_lblcr.c 	if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) {
atomic_read       631 net/netfilter/ipvs/ip_vs_lblcr.c 			if (atomic_read(&d->activeconns)*2
atomic_read       632 net/netfilter/ipvs/ip_vs_lblcr.c 			    < atomic_read(&d->weight)) {
atomic_read       663 net/netfilter/ipvs/ip_vs_lblcr.c 		if (atomic_read(&en->set.size) > 1 &&
atomic_read       667 net/netfilter/ipvs/ip_vs_lblcr.c 			if (atomic_read(&en->set.size) > 1) {
atomic_read        43 net/netfilter/ipvs/ip_vs_lc.c 		    atomic_read(&dest->weight) == 0)
atomic_read        59 net/netfilter/ipvs/ip_vs_lc.c 			      atomic_read(&least->activeconns),
atomic_read        60 net/netfilter/ipvs/ip_vs_lc.c 			      atomic_read(&least->inactconns));
atomic_read        83 net/netfilter/ipvs/ip_vs_mh.c 	return atomic_read(&dest->weight) <= 0 ||
atomic_read       151 net/netfilter/ipvs/ip_vs_mh.c 		lw = atomic_read(&dest->last_weight);
atomic_read       330 net/netfilter/ipvs/ip_vs_mh.c 		weight = atomic_read(&dest->last_weight);
atomic_read       358 net/netfilter/ipvs/ip_vs_mh.c 		new_weight = atomic_read(&dest->last_weight);
atomic_read        45 net/netfilter/ipvs/ip_vs_nq.c 	return atomic_read(&dest->activeconns) + 1;
atomic_read        77 net/netfilter/ipvs/ip_vs_nq.c 		    !atomic_read(&dest->weight))
atomic_read        83 net/netfilter/ipvs/ip_vs_nq.c 		if (atomic_read(&dest->activeconns) == 0) {
atomic_read        90 net/netfilter/ipvs/ip_vs_nq.c 		    ((__s64)loh * atomic_read(&dest->weight) >
atomic_read        91 net/netfilter/ipvs/ip_vs_nq.c 		     (__s64)doh * atomic_read(&least->weight))) {
atomic_read       107 net/netfilter/ipvs/ip_vs_nq.c 		      atomic_read(&least->activeconns),
atomic_read       109 net/netfilter/ipvs/ip_vs_nq.c 		      atomic_read(&least->weight), loh);
atomic_read        36 net/netfilter/ipvs/ip_vs_ovf.c 		w = atomic_read(&dest->weight);
atomic_read        38 net/netfilter/ipvs/ip_vs_ovf.c 		    atomic_read(&dest->activeconns) > w ||
atomic_read        51 net/netfilter/ipvs/ip_vs_ovf.c 			      atomic_read(&h->activeconns),
atomic_read        52 net/netfilter/ipvs/ip_vs_ovf.c 			      atomic_read(&h->weight));
atomic_read        71 net/netfilter/ipvs/ip_vs_rr.c 			    atomic_read(&dest->weight) > 0)
atomic_read        94 net/netfilter/ipvs/ip_vs_rr.c 		      atomic_read(&dest->activeconns),
atomic_read        95 net/netfilter/ipvs/ip_vs_rr.c 		      refcount_read(&dest->refcnt), atomic_read(&dest->weight));
atomic_read        49 net/netfilter/ipvs/ip_vs_sed.c 	return atomic_read(&dest->activeconns) + 1;
atomic_read        80 net/netfilter/ipvs/ip_vs_sed.c 		    atomic_read(&dest->weight) > 0) {
atomic_read        97 net/netfilter/ipvs/ip_vs_sed.c 		if ((__s64)loh * atomic_read(&dest->weight) >
atomic_read        98 net/netfilter/ipvs/ip_vs_sed.c 		    (__s64)doh * atomic_read(&least->weight)) {
atomic_read       108 net/netfilter/ipvs/ip_vs_sed.c 		      atomic_read(&least->activeconns),
atomic_read       110 net/netfilter/ipvs/ip_vs_sed.c 		      atomic_read(&least->weight), loh);
atomic_read        76 net/netfilter/ipvs/ip_vs_sh.c 	return atomic_read(&dest->weight) <= 0 ||
atomic_read       192 net/netfilter/ipvs/ip_vs_sh.c 				      atomic_read(&dest->weight));
atomic_read       195 net/netfilter/ipvs/ip_vs_sh.c 			if (++d_count >= atomic_read(&dest->weight)) {
atomic_read        52 net/netfilter/ipvs/ip_vs_wlc.c 		    atomic_read(&dest->weight) > 0) {
atomic_read        69 net/netfilter/ipvs/ip_vs_wlc.c 		if ((__s64)loh * atomic_read(&dest->weight) >
atomic_read        70 net/netfilter/ipvs/ip_vs_wlc.c 		    (__s64)doh * atomic_read(&least->weight)) {
atomic_read        80 net/netfilter/ipvs/ip_vs_wlc.c 		      atomic_read(&least->activeconns),
atomic_read        82 net/netfilter/ipvs/ip_vs_wlc.c 		      atomic_read(&least->weight), loh);
atomic_read        76 net/netfilter/ipvs/ip_vs_wrr.c 		weight = atomic_read(&dest->weight);
atomic_read        97 net/netfilter/ipvs/ip_vs_wrr.c 		new_weight = atomic_read(&dest->weight);
atomic_read       181 net/netfilter/ipvs/ip_vs_wrr.c 			    atomic_read(&dest->weight) >= mark->cw)
atomic_read       215 net/netfilter/ipvs/ip_vs_wrr.c 		      atomic_read(&dest->activeconns),
atomic_read       217 net/netfilter/ipvs/ip_vs_wrr.c 		      atomic_read(&dest->weight));
atomic_read       342 net/netfilter/ipvs/ip_vs_xmit.c 				  atomic_read(&rt->dst.__refcnt));
atomic_read       513 net/netfilter/ipvs/ip_vs_xmit.c 				  atomic_read(&rt->dst.__refcnt));
atomic_read       601 net/netfilter/nf_conntrack_core.c 	WARN_ON(atomic_read(&nfct->use) != 0);
atomic_read      1272 net/netfilter/nf_conntrack_core.c 			if (atomic_read(&net->ct.count) < nf_conntrack_max95)
atomic_read      1357 net/netfilter/nf_conntrack_core.c 	    unlikely(atomic_read(&net->ct.count) > nf_conntrack_max)) {
atomic_read      1417 net/netfilter/nf_conntrack_core.c 	WARN_ON(atomic_read(&ct->ct_general.use) != 0);
atomic_read      2164 net/netfilter/nf_conntrack_core.c 	if (atomic_read(&net->ct.count) > 0) {
atomic_read      2180 net/netfilter/nf_conntrack_core.c 	if (atomic_read(&net->ct.count) == 0)
atomic_read      2209 net/netfilter/nf_conntrack_core.c 		if (atomic_read(&net->ct.count) == 0)
atomic_read      2290 net/netfilter/nf_conntrack_core.c 		if (atomic_read(&net->ct.count) != 0)
atomic_read       501 net/netfilter/nf_conntrack_netlink.c 	if (nla_put_be32(skb, CTA_USE, htonl(atomic_read(&ct->ct_general.use))))
atomic_read      2289 net/netfilter/nf_conntrack_netlink.c 	unsigned int nr_conntracks = atomic_read(&net->ct.count);
atomic_read       367 net/netfilter/nf_conntrack_standalone.c 	seq_printf(s, "use=%u\n", atomic_read(&ct->ct_general.use));
atomic_read       425 net/netfilter/nf_conntrack_standalone.c 	unsigned int nr_conntracks = atomic_read(&net->ct.count);
atomic_read       239 net/netfilter/nf_nat_masquerade.c 	if (event != NETDEV_DOWN || atomic_read(&v6_worker_count) >= 16)
atomic_read      3821 net/netfilter/nf_tables_api.c 	    (nlh->nlmsg_flags & NLM_F_NONREC && atomic_read(&set->nelems) > 0)) {
atomic_read       263 net/netfilter/nft_ct.c 	if (likely(atomic_read(&ct->ct_general.use) == 1)) {
atomic_read        30 net/netfilter/nft_numgen.c 		oval = atomic_read(&priv->counter);
atomic_read        42 net/netfilter/xt_statistic.c 			oval = atomic_read(&info->master->count);
atomic_read       957 net/netlabel/netlabel_kapi.c 	return (atomic_read(&netlabel_mgmt_protocount) > 0);
atomic_read       401 net/netlink/af_netlink.c 	WARN_ON(atomic_read(&sk->sk_rmem_alloc));
atomic_read       427 net/netlink/af_netlink.c 	if (atomic_read(&nl_table_users)) {
atomic_read       433 net/netlink/af_netlink.c 			if (atomic_read(&nl_table_users) == 0)
atomic_read      1210 net/netlink/af_netlink.c 	if ((atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf ||
atomic_read      1224 net/netlink/af_netlink.c 		if ((atomic_read(&sk->sk_rmem_alloc) > sk->sk_rcvbuf ||
atomic_read      1379 net/netlink/af_netlink.c 	if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf &&
atomic_read      1383 net/netlink/af_netlink.c 		return atomic_read(&sk->sk_rmem_alloc) > (sk->sk_rcvbuf >> 1);
atomic_read      2000 net/netlink/af_netlink.c 	    atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf / 2) {
atomic_read      2206 net/netlink/af_netlink.c 	if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf)
atomic_read      2637 net/netlink/af_netlink.c 			   atomic_read(&s->sk_drops),
atomic_read       419 net/netlink/genetlink.c 		   atomic_read(&genl_sk_destructing_cnt) == 0);
atomic_read       136 net/netrom/nr_timer.c 		if (atomic_read(&sk->sk_rmem_alloc) < (sk->sk_rcvbuf / 2) &&
atomic_read       463 net/nfc/llcp_core.c 		pr_debug("%d clients\n", atomic_read(client_cnt));
atomic_read       786 net/nfc/nci/core.c 	if ((atomic_read(&ndev->state) == NCI_DISCOVERY) ||
atomic_read       787 net/nfc/nci/core.c 	    (atomic_read(&ndev->state) == NCI_W4_ALL_DISCOVERIES)) {
atomic_read       797 net/nfc/nci/core.c 	if ((atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) ||
atomic_read       798 net/nfc/nci/core.c 	    (atomic_read(&ndev->state) == NCI_POLL_ACTIVE)) {
atomic_read       837 net/nfc/nci/core.c 	if ((atomic_read(&ndev->state) != NCI_DISCOVERY) &&
atomic_read       838 net/nfc/nci/core.c 	    (atomic_read(&ndev->state) != NCI_W4_ALL_DISCOVERIES)) {
atomic_read       858 net/nfc/nci/core.c 	if ((atomic_read(&ndev->state) != NCI_W4_HOST_SELECT) &&
atomic_read       859 net/nfc/nci/core.c 	    (atomic_read(&ndev->state) != NCI_POLL_ACTIVE)) {
atomic_read       887 net/nfc/nci/core.c 	if (atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) {
atomic_read       935 net/nfc/nci/core.c 	if (atomic_read(&ndev->state) == NCI_POLL_ACTIVE) {
atomic_read       972 net/nfc/nci/core.c 		if (atomic_read(&ndev->state) == NCI_LISTEN_ACTIVE ||
atomic_read       973 net/nfc/nci/core.c 		    atomic_read(&ndev->state) == NCI_DISCOVERY) {
atomic_read      1432 net/nfc/nci/core.c 	pr_debug("credits_cnt %d\n", atomic_read(&conn_info->credits_cnt));
atomic_read      1435 net/nfc/nci/core.c 	while (atomic_read(&conn_info->credits_cnt)) {
atomic_read      1441 net/nfc/nci/core.c 		if (atomic_read(&conn_info->credits_cnt) !=
atomic_read      1510 net/nfc/nci/core.c 	pr_debug("cmd_cnt %d\n", atomic_read(&ndev->cmd_cnt));
atomic_read      1513 net/nfc/nci/core.c 	if (atomic_read(&ndev->cmd_cnt)) {
atomic_read        72 net/nfc/nci/ntf.c 	if (atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) {
atomic_read       637 net/nfc/nci/ntf.c 		if (atomic_read(&ndev->state) == NCI_DISCOVERY) {
atomic_read       180 net/nfc/nci/rsp.c 	    (atomic_read(&ndev->state) != NCI_POLL_ACTIVE)) {
atomic_read       661 net/packet/af_packet.c 		while (atomic_read(&pkc->blk_fill_in_prog)) {
atomic_read       762 net/packet/af_packet.c 	if (atomic_read(&po->tp_drops))
atomic_read       923 net/packet/af_packet.c 			while (atomic_read(&pkc->blk_fill_in_prog)) {
atomic_read      1236 net/packet/af_packet.c 		int avail = rcvbuf - atomic_read(&sk->sk_rmem_alloc)
atomic_read      1286 net/packet/af_packet.c 	WARN_ON(atomic_read(&sk->sk_rmem_alloc));
atomic_read      2090 net/packet/af_packet.c 	if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf)
atomic_read      2242 net/packet/af_packet.c 			    atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf) {
atomic_read      2299 net/packet/af_packet.c 		if (atomic_read(&po->tp_drops))
atomic_read      4304 net/packet/af_packet.c 		if (atomic_read(&po->mapped))
atomic_read      4407 net/packet/af_packet.c 	if (closing || atomic_read(&po->mapped) == 0) {
atomic_read      4425 net/packet/af_packet.c 		if (atomic_read(&po->mapped))
atomic_read      4427 net/packet/af_packet.c 			       atomic_read(&po->mapped));
atomic_read      4609 net/packet/af_packet.c 			   atomic_read(&s->sk_rmem_alloc),
atomic_read      1158 net/phonet/pep.c 	done = atomic_read(&pn->tx_credits);
atomic_read      1172 net/phonet/pep.c 		done = sk_wait_event(sk, &timeo, atomic_read(&pn->tx_credits), &wait);
atomic_read      1194 net/phonet/pep.c 	return atomic_read(&pn->tx_credits);
atomic_read       350 net/phonet/socket.c 		atomic_read(&pn->tx_credits))
atomic_read       604 net/phonet/socket.c 			atomic_read(&sk->sk_drops));
atomic_read       284 net/rds/cong.c 	unsigned long gen = atomic_read(&rds_cong_generation);
atomic_read       375 net/rds/connection.c 					    atomic_read(&cp->cp_state));
atomic_read       407 net/rds/connection.c 					    atomic_read(&cp->cp_state));
atomic_read       748 net/rds/connection.c 			  atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING,
atomic_read       751 net/rds/connection.c 			  atomic_read(&cp->cp_state) == RDS_CONN_UP,
atomic_read       774 net/rds/connection.c 			  atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING,
atomic_read       777 net/rds/connection.c 			  atomic_read(&cp->cp_state) == RDS_CONN_UP,
atomic_read       322 net/rds/ib.c   		iinfo->cache_allocs = atomic_read(&ic->i_cache_allocs);
atomic_read       359 net/rds/ib.c   		iinfo6->cache_allocs = atomic_read(&ic->i_cache_allocs);
atomic_read       505 net/rds/ib.c   		atomic_read(&rds_ib_unloading) != 0);
atomic_read       255 net/rds/ib_cm.c 				(atomic_read(&ic->i_credits));
atomic_read       323 net/rds/ib_cm.c 	if (atomic_read(&ic->i_cq_quiesce))
atomic_read       368 net/rds/ib_cm.c 	if (atomic_read(&ic->i_cq_quiesce))
atomic_read      1021 net/rds/ib_cm.c 			   (atomic_read(&ic->i_signaled_sends) == 0) &&
atomic_read      1022 net/rds/ib_cm.c 			   (atomic_read(&ic->i_fastreg_inuse_count) == 0) &&
atomic_read      1023 net/rds/ib_cm.c 			   (atomic_read(&ic->i_fastreg_wrs) == RDS_IB_DEFAULT_FR_WR));
atomic_read        47 net/rds/ib_fmr.c 	if (atomic_read(&pool->dirty_count) >= pool->max_items / 10)
atomic_read        51 net/rds/ib_fmr.c 	if (atomic_read(&pool->dirty_count) >=  pool->max_items * 9 / 10) {
atomic_read        92 net/rds/ib_frmr.c 	if (atomic_read(&pool->item_count) > pool->max_items_soft)
atomic_read       118 net/rds/ib_frmr.c 	if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned ||
atomic_read       119 net/rds/ib_frmr.c 	    atomic_read(&pool->dirty_count) >= pool->max_items / 5)
atomic_read       275 net/rds/ib_rdma.c 	item_count = atomic_read(&pool->item_count);
atomic_read       495 net/rds/ib_rdma.c 	if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned ||
atomic_read       496 net/rds/ib_rdma.c 	    atomic_read(&pool->dirty_count) >= pool->max_items / 5)
atomic_read       574 net/rds/ib_rdma.c 	WARN_ON(atomic_read(&pool->item_count));
atomic_read       575 net/rds/ib_rdma.c 	WARN_ON(atomic_read(&pool->free_pinned));
atomic_read      1076 net/rds/ib_recv.c 	WARN_ON(atomic_read(&rds_ib_allocation));
atomic_read        78 net/rds/ib_ring.c 	diff = ring->w_alloc_ctr - (u32) atomic_read(&ring->w_free_ctr);
atomic_read       232 net/rds/ib_send.c 	BUG_ON(atomic_read(&ic->i_signaled_sends) < 0);
atomic_read       366 net/rds/ib_send.c 	oldval = newval = atomic_read(&ic->i_credits);
atomic_read       416 net/rds/ib_send.c 			IB_GET_SEND_CREDITS(atomic_read(&ic->i_credits)),
atomic_read       449 net/rds/ib_send.c 	if (IB_GET_POST_CREDITS(atomic_read(&ic->i_credits)) >= 16)
atomic_read        55 net/rds/loop.c 	return atomic_read(&rds_loop_unloading) != 0;
atomic_read       814 net/rds/rds.h  	return atomic_read(&cp->cp_state);
atomic_read       827 net/rds/rds.h  	return atomic_read(&cp->cp_state) == RDS_CONN_UP;
atomic_read       840 net/rds/rds.h  	return atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING;
atomic_read       434 net/rds/tcp.c  	return atomic_read(&rds_tcp_unloading) != 0;
atomic_read        80 net/rds/threads.c 				atomic_read(&cp->cp_state));
atomic_read       418 net/rfkill/core.c 	if (atomic_read(&rfkill_input_disabled))
atomic_read       443 net/rfkill/core.c 	if (atomic_read(&rfkill_input_disabled))
atomic_read       471 net/rfkill/core.c 	if (atomic_read(&rfkill_input_disabled))
atomic_read       490 net/rfkill/core.c 	if (atomic_read(&rfkill_input_disabled))
atomic_read      1051 net/rfkill/core.c 		if (!atomic_read(&rfkill_input_disabled))
atomic_read       181 net/rose/rose_in.c 			if (atomic_read(&sk->sk_rmem_alloc) >
atomic_read       141 net/rose/rose_timer.c 		if (atomic_read(&sk->sk_rmem_alloc) < (sk->sk_rcvbuf / 2) &&
atomic_read       353 net/rxrpc/af_rxrpc.c 	_enter("%d{%d}", call->debug_id, atomic_read(&call->usage));
atomic_read      1047 net/rxrpc/af_rxrpc.c 	ASSERTCMP(atomic_read(&rxrpc_n_tx_skbs), ==, 0);
atomic_read      1048 net/rxrpc/af_rxrpc.c 	ASSERTCMP(atomic_read(&rxrpc_n_rx_skbs), ==, 0);
atomic_read        19 net/rxrpc/ar-internal.h 	BUG_ON(atomic_read((X)) >> (sizeof(atomic_t) - 2) == \
atomic_read        88 net/rxrpc/call_accept.c 				 atomic_read(&conn->usage), here);
atomic_read       101 net/rxrpc/call_accept.c 			 atomic_read(&call->usage),
atomic_read        91 net/rxrpc/call_object.c 	_leave(" = %p [%d]", call, atomic_read(&call->usage));
atomic_read       243 net/rxrpc/call_object.c 			 atomic_read(&call->usage),
atomic_read       294 net/rxrpc/call_object.c 			 atomic_read(&call->usage), here, NULL);
atomic_read       317 net/rxrpc/call_object.c 			 atomic_read(&call->usage), here, ERR_PTR(ret));
atomic_read       392 net/rxrpc/call_object.c 	int n = atomic_read(&call->usage);
atomic_read       409 net/rxrpc/call_object.c 		int n = atomic_read(&call->usage);
atomic_read       449 net/rxrpc/call_object.c 	_enter("{%d,%d}", call->debug_id, atomic_read(&call->usage));
atomic_read       452 net/rxrpc/call_object.c 			 atomic_read(&call->usage),
atomic_read       639 net/rxrpc/call_object.c 			       call, atomic_read(&call->usage),
atomic_read       652 net/rxrpc/call_object.c 	wait_var_event(&rxnet->nr_calls, !atomic_read(&rxnet->nr_calls));
atomic_read       156 net/rxrpc/conn_client.c 			       conn, atomic_read(&conn->usage));
atomic_read       216 net/rxrpc/conn_client.c 			 atomic_read(&conn->usage),
atomic_read       110 net/rxrpc/conn_object.c 		if (!conn || atomic_read(&conn->usage) == 0)
atomic_read       120 net/rxrpc/conn_object.c 		if (!conn || atomic_read(&conn->usage) == 0) {
atomic_read       284 net/rxrpc/conn_object.c 		int n = atomic_read(&conn->usage);
atomic_read       354 net/rxrpc/conn_object.c 	_enter("{%d,u=%d}", conn->debug_id, atomic_read(&conn->usage));
atomic_read       356 net/rxrpc/conn_object.c 	ASSERTCMP(atomic_read(&conn->usage), ==, 0);
atomic_read       395 net/rxrpc/conn_object.c 		ASSERTCMP(atomic_read(&conn->usage), >, 0);
atomic_read       396 net/rxrpc/conn_object.c 		if (likely(atomic_read(&conn->usage) > 1))
atomic_read       408 net/rxrpc/conn_object.c 			       conn->debug_id, atomic_read(&conn->usage),
atomic_read       445 net/rxrpc/conn_object.c 		ASSERTCMP(atomic_read(&conn->usage), ==, 0);
atomic_read       473 net/rxrpc/conn_object.c 		       conn, atomic_read(&conn->usage));
atomic_read       484 net/rxrpc/conn_object.c 	wait_var_event(&rxnet->nr_conns, !atomic_read(&rxnet->nr_conns));
atomic_read        96 net/rxrpc/conn_service.c 	if (atomic_read(&cursor->usage) == 0)
atomic_read       138 net/rxrpc/conn_service.c 				 atomic_read(&conn->usage),
atomic_read      1403 net/rxrpc/input.c 	if (!call || atomic_read(&call->usage) == 0) {
atomic_read       344 net/rxrpc/local_object.c 	int n = atomic_read(&local->usage);
atomic_read       452 net/rxrpc/local_object.c 			  atomic_read(&local->usage), NULL);
atomic_read       508 net/rxrpc/local_object.c 			       local, atomic_read(&local->usage));
atomic_read       124 net/rxrpc/peer_object.c 		    atomic_read(&peer->usage) > 0)
atomic_read       143 net/rxrpc/peer_object.c 		_leave(" = %p {u=%d}", peer, atomic_read(&peer->usage));
atomic_read       373 net/rxrpc/peer_object.c 	_leave(" = %p {u=%d}", peer, atomic_read(&peer->usage));
atomic_read       478 net/rxrpc/peer_object.c 			       atomic_read(&peer->usage),
atomic_read       110 net/rxrpc/proc.c 		   atomic_read(&call->usage),
atomic_read       192 net/rxrpc/proc.c 		   atomic_read(&conn->usage),
atomic_read       195 net/rxrpc/proc.c 		   atomic_read(&conn->serial),
atomic_read       242 net/rxrpc/proc.c 		   atomic_read(&peer->usage),
atomic_read        38 net/rxrpc/skbuff.c 		int n = atomic_read(select_skb_count(skb));
atomic_read       162 net/sched/act_api.c 		if (!bind && strict && atomic_read(&p->tcfa_bindcnt) > 0)
atomic_read       275 net/sched/act_api.c 	if (atomic_read(&p->tcfa_bindcnt) > 0)
atomic_read       379 net/sched/act_api.c 	if (!atomic_read(&p->tcfa_bindcnt)) {
atomic_read       142 net/sched/act_bpf.c 		.bindcnt = atomic_read(&prog->tcf_bindcnt) - bind,
atomic_read       177 net/sched/act_connmark.c 		.bindcnt = atomic_read(&ci->tcf_bindcnt) - bind,
atomic_read       641 net/sched/act_csum.c 		.bindcnt = atomic_read(&p->tcf_bindcnt) - bind,
atomic_read       839 net/sched/act_ct.c 		.bindcnt = atomic_read(&c->tcf_bindcnt) - bind,
atomic_read       289 net/sched/act_ctinfo.c 		.bindcnt = atomic_read(&ci->tcf_bindcnt) - bind,
atomic_read       200 net/sched/act_gact.c 		.bindcnt = atomic_read(&gact->tcf_bindcnt) - bind,
atomic_read       627 net/sched/act_ife.c 		.bindcnt = atomic_read(&ife->tcf_bindcnt) - bind,
atomic_read       294 net/sched/act_ipt.c 	c.bindcnt = atomic_read(&ipt->tcf_bindcnt) - bind;
atomic_read       336 net/sched/act_mirred.c 		.bindcnt = atomic_read(&m->tcf_bindcnt) - bind,
atomic_read       306 net/sched/act_mpls.c 		.bindcnt  = atomic_read(&m->tcf_bindcnt) - bind,
atomic_read       270 net/sched/act_nat.c 		.bindcnt  = atomic_read(&p->tcf_bindcnt) - bind,
atomic_read       435 net/sched/act_pedit.c 	opt->bindcnt = atomic_read(&p->tcf_bindcnt) - bind;
atomic_read       313 net/sched/act_police.c 		.bindcnt = atomic_read(&police->tcf_bindcnt) - bind,
atomic_read       207 net/sched/act_sample.c 		.bindcnt    = atomic_read(&s->tcf_bindcnt) - bind,
atomic_read       178 net/sched/act_simple.c 		.bindcnt = atomic_read(&d->tcf_bindcnt) - bind,
atomic_read       237 net/sched/act_skbedit.c 		.bindcnt = atomic_read(&d->tcf_bindcnt) - bind,
atomic_read       223 net/sched/act_skbmod.c 		.bindcnt = atomic_read(&d->tcf_bindcnt) - bind,
atomic_read       521 net/sched/act_tunnel_key.c 		.bindcnt  = atomic_read(&t->tcf_bindcnt) - bind,
atomic_read       261 net/sched/act_vlan.c 		.bindcnt  = atomic_read(&v->tcf_bindcnt) - bind,
atomic_read       702 net/sched/cls_api.c 	return atomic_read(&block->offloadcnt);
atomic_read       416 net/sched/em_meta.c 	dst->value = atomic_read(&sk->sk_omem_alloc);
atomic_read       413 net/sctp/associola.c 	WARN_ON(atomic_read(&asoc->rmem_alloc));
atomic_read      1436 net/sctp/associola.c 						   atomic_read(&t->mtu_info));
atomic_read      1538 net/sctp/associola.c 		rx_count = atomic_read(&asoc->rmem_alloc);
atomic_read      1540 net/sctp/associola.c 		rx_count = atomic_read(&asoc->base.sk->sk_rmem_alloc);
atomic_read       169 net/sctp/diag.c 			amt = atomic_read(&asoc->rmem_alloc);
atomic_read       177 net/sctp/diag.c 		mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc);
atomic_read       179 net/sctp/diag.c 		mem[SK_MEMINFO_DROPS] = atomic_read(&sk->sk_drops);
atomic_read       421 net/sctp/diag.c 		r->idiag_rqueue = atomic_read(&infox->asoc->rmem_alloc);
atomic_read        69 net/sctp/objcnt.c 				atomic_read(sctp_dbg_objcnt[i].counter));
atomic_read       259 net/sctp/proc.c 		   atomic_read(&assoc->rmem_alloc),
atomic_read      9432 net/sctp/socket.c 	if (atomic_read(&sctp_sk(oldsk)->pd_mode)) {
atomic_read       633 net/sctp/ulpevent.c 		rx_count = atomic_read(&asoc->rmem_alloc);
atomic_read       635 net/sctp/ulpevent.c 		rx_count = atomic_read(&sk->sk_rmem_alloc);
atomic_read       215 net/sctp/ulpqueue.c 	if (atomic_read(&sp->pd_mode) == 0) {
atomic_read       483 net/sctp/ulpqueue.c 		    atomic_read(&sctp_sk(asoc->base.sk)->pd_mode))
atomic_read      1059 net/sctp/ulpqueue.c 	if (sp->frag_interleave || atomic_read(&sp->pd_mode) == 0) {
atomic_read      1494 net/smc/af_smc.c 		} else if (!atomic_read(&smc_sk(nsk)->conn.bytes_to_rcv)) {
atomic_read      1633 net/smc/af_smc.c 			     atomic_read(&smc->conn.sndbuf_space)) ||
atomic_read      1640 net/smc/af_smc.c 			if (atomic_read(&smc->conn.bytes_to_rcv))
atomic_read      1817 net/smc/af_smc.c 			answ = atomic_read(&smc->conn.bytes_to_rcv);
atomic_read      1830 net/smc/af_smc.c 					atomic_read(&smc->conn.sndbuf_space);
atomic_read       287 net/smc/smc_cdc.c 	    atomic_read(&conn->peer_rmbe_space) == conn->peer_rmbe_size) {
atomic_read        94 net/smc/smc_close.c 	if (atomic_read(&conn->bytes_to_rcv))
atomic_read       187 net/smc/smc_rx.c 	return atomic_read(&conn->bytes_to_rcv) &&
atomic_read       188 net/smc/smc_rx.c 	       !atomic_read(&conn->splice_pending);
atomic_read       367 net/smc/smc_rx.c 		readable = atomic_read(&conn->bytes_to_rcv);
atomic_read       368 net/smc/smc_rx.c 		splbytes = atomic_read(&conn->splice_pending);
atomic_read        28 net/smc/smc_rx.h 	return atomic_read(&conn->bytes_to_rcv);
atomic_read        47 net/smc/smc_tx.c 	if (atomic_read(&smc->conn.sndbuf_space) && sock) {
atomic_read       108 net/smc/smc_tx.c 		if (atomic_read(&conn->sndbuf_space) && !conn->urg_tx_pend)
atomic_read       115 net/smc/smc_tx.c 			      (atomic_read(&conn->sndbuf_space) &&
atomic_read       166 net/smc/smc_tx.c 		if (!atomic_read(&conn->sndbuf_space) || conn->urg_tx_pend) {
atomic_read       177 net/smc/smc_tx.c 		writespace = atomic_read(&conn->sndbuf_space);
atomic_read       225 net/smc/smc_tx.c 		    (atomic_read(&conn->sndbuf_space) >
atomic_read       421 net/smc/smc_tx.c 	rmbespace = atomic_read(&conn->peer_rmbe_space);
atomic_read       583 net/smc/smc_tx.c 	    !atomic_read(&conn->bytes_to_rcv))
atomic_read       619 net/smc/smc_tx.c 	    !atomic_read(&conn->bytes_to_rcv))
atomic_read       605 net/sunrpc/auth_gss/gss_krb5_mech.c 	if (seq_send64 != atomic_read(&ctx->seq_send)) {
atomic_read       607 net/sunrpc/auth_gss/gss_krb5_mech.c 			seq_send64, atomic_read(&ctx->seq_send));
atomic_read       249 net/sunrpc/backchannel_rqst.c 		if (atomic_read(&xprt->bc_slot_count) >= BC_MAX_SLOTS)
atomic_read      1172 net/sunrpc/cache.c 	if (atomic_read(&detail->writers))
atomic_read       173 net/sunrpc/clnt.c 		if (atomic_read(&clnt->cl_count) == 0)
atomic_read      1083 net/sunrpc/clnt.c 		if (atomic_read(&clnt->cl_swapper))
atomic_read      1226 net/sunrpc/clnt.c 	WARN_ON_ONCE(atomic_read(&task->tk_count) != 2);
atomic_read      1211 net/sunrpc/sched.c 	if (atomic_read(&task->tk_count) != 1 + !RPC_IS_ASYNC(task)) {
atomic_read      1613 net/sunrpc/svc.c 	WARN_ON_ONCE(atomic_read(&task->tk_count) != 1);
atomic_read       344 net/sunrpc/svc_xprt.c 	int nrqsts = atomic_read(&xprt->xpt_nr_rqsts);
atomic_read       629 net/sunrpc/svcsock.c 	required = atomic_read(&svsk->sk_xprt.xpt_reserved) + serv->sv_max_mesg;
atomic_read       986 net/sunrpc/xprt.c 	return atomic_read(&req->rq_pin) != 0;
atomic_read        95 net/sunrpc/xprtrdma/svc_rdma.c 		int len = snprintf(str_buf, 32, "%d\n", atomic_read(stat));
atomic_read       334 net/sunrpc/xprtrdma/svc_rdma_rw.c 			   atomic_read(&rdma->sc_sq_avail) > cc->cc_sqecount);
atomic_read       301 net/sunrpc/xprtrdma/svc_rdma_sendto.c 				   atomic_read(&rdma->sc_sq_avail) > 1);
atomic_read      1282 net/sunrpc/xprtsock.c 	if (atomic_read(&transport->xprt.swapper))
atomic_read      2067 net/sunrpc/xprtsock.c 	if (atomic_read(&xprt->swapper))
atomic_read      2268 net/tipc/socket.c 		lim = rcvbuf_limit(sk, skb) + atomic_read(dcnt);
atomic_read      3473 net/tipc/socket.c 			atomic_read(&sk->sk_drops)))
atomic_read      3718 net/tipc/socket.c 	unsigned int lim = rcvbuf_limit(sk, skb) + atomic_read(dcnt);
atomic_read      3794 net/tipc/socket.c 	i += scnprintf(buf + i, sz - i, " %u", atomic_read(&tsk->dupl_rcvcnt));
atomic_read       371 net/tipc/topsrv.c 	if (atomic_read(&tn->subscription_count) >= TIPC_MAX_SUBSCR) {
atomic_read      1103 net/tls/tls_sw.c 		pending = atomic_read(&ctx->encrypt_pending);
atomic_read      1906 net/tls/tls_sw.c 		pending = atomic_read(&ctx->decrypt_pending);
atomic_read      2127 net/tls/tls_sw.c 	if (atomic_read(&ctx->encrypt_pending))
atomic_read       368 net/vmw_vsock/virtio_transport.c 	val = atomic_read(&vsock->queued_replies);
atomic_read        96 net/wireless/lib80211.c 		if (atomic_read(&entry->refcnt) != 0 && !force)
atomic_read        81 net/wireless/scan.c 	if (WARN_ON(atomic_read(&bss->hold)))
atomic_read       404 net/wireless/scan.c 		if (atomic_read(&bss->hold))
atomic_read       425 net/wireless/scan.c 		if (atomic_read(&bss->hold))
atomic_read       945 net/wireless/scan.c 		    !atomic_read(&bss->hold))
atomic_read      2071 net/wireless/scan.c 		WARN_ON(atomic_read(&new->hold));
atomic_read       291 net/x25/x25_in.c 				if (atomic_read(&sk->sk_rmem_alloc) >
atomic_read       370 net/x25/x25_subr.c 	if (atomic_read(&sk->sk_rmem_alloc) < (sk->sk_rcvbuf >> 1) &&
atomic_read      2744 net/xfrm/xfrm_policy.c 	xdst->policy_genid = atomic_read(&pols[0]->genid);
atomic_read      3809 net/xfrm/xfrm_policy.c 		    xdst->policy_genid != atomic_read(&xdst->pols[0]->genid))
atomic_read      2401 net/xfrm/xfrm_state.c 		if (atomic_read(&t->tunnel_users) == 2)
atomic_read       118 security/integrity/ima/ima_main.c 		if (atomic_read(&inode->i_readcount) && IS_IMA(inode)) {
atomic_read       156 security/integrity/ima/ima_main.c 	if (atomic_read(&inode->i_writecount) == 1) {
atomic_read       315 security/keys/proc.c 		   atomic_read(&user->nkeys),
atomic_read       316 security/keys/proc.c 		   atomic_read(&user->nikeys),
atomic_read       176 security/selinux/avc.c 			 atomic_read(&avc->avc_cache.active_nodes),
atomic_read       165 security/selinux/hooks.c 		atomic_read(&selinux_secmark_refcount));
atomic_read        36 security/selinux/include/xfrm.h 	return (atomic_read(&selinux_xfrm_refcount) > 0);
atomic_read      2083 security/tomoyo/common.c 		if (atomic_read(&tomoyo_query_observers))
atomic_read      2127 security/tomoyo/common.c 		     !atomic_read(&tomoyo_query_observers), HZ))
atomic_read      2360 security/tomoyo/common.c 				 atomic_read(&tomoyo_stat_updated[i]));
atomic_read       420 security/tomoyo/condition.c 		    atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS)
atomic_read       451 security/tomoyo/gc.c 		if (atomic_read(&container_of
atomic_read       538 security/tomoyo/gc.c 			if (!domain->is_deleted || atomic_read(&domain->users))
atomic_read       555 security/tomoyo/gc.c 			if (atomic_read(&ptr->users) > 0)
atomic_read       581 security/tomoyo/gc.c 				    atomic_read(&group->head.users) > 0)
atomic_read       596 security/tomoyo/gc.c 			if (atomic_read(&ptr->users) > 0)
atomic_read       114 security/tomoyo/memory.c 		    atomic_read(&group->head.users) == TOMOYO_GC_IN_PROGRESS)
atomic_read       168 security/tomoyo/memory.c 		    atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS)
atomic_read       697 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count)) {
atomic_read       816 sound/core/oss/pcm_oss.c 	if (atomic_read(&runtime->oss.rw_ref)) {
atomic_read       853 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read       862 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read       989 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count) ||
atomic_read       998 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count) ||
atomic_read      1374 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read      1484 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read      1639 sound/core/oss/pcm_oss.c 		if (atomic_read(&substream->mmap_count))
atomic_read      1818 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read      2057 sound/core/oss/pcm_oss.c 			if (atomic_read(&psubstream->mmap_count))
atomic_read      2185 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count)) {
atomic_read      2771 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read      2782 sound/core/oss/pcm_oss.c 	if (atomic_read(&substream->mmap_count))
atomic_read       658 sound/core/pcm_native.c 		if (atomic_read(&substream->mmap_count))
atomic_read       781 sound/core/pcm_native.c 	if (atomic_read(&substream->mmap_count))
atomic_read       423 sound/core/seq/seq_clientmgr.c 	if (atomic_read(&fifo->overflow) > 0) {
atomic_read       686 sound/core/seq/seq_clientmgr.c 		if (atomic_read(&subs->ref_count) != 2)
atomic_read      1863 sound/core/seq/seq_clientmgr.c 			if (atomic_read(&client->pool->counter))
atomic_read        16 sound/core/seq/seq_lock.c 	if (atomic_read(lockp) < 0) {
atomic_read        17 sound/core/seq/seq_lock.c 		pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line);
atomic_read        20 sound/core/seq/seq_lock.c 	while (atomic_read(lockp) > 0) {
atomic_read        22 sound/core/seq/seq_lock.c 			pr_warn("ALSA: seq_lock: waiting [%d left] in %s:%d\n", atomic_read(lockp), file, line);
atomic_read        24 sound/core/seq/seq_memory.c 	return pool->total_elements - atomic_read(&pool->counter);
atomic_read       257 sound/core/seq/seq_memory.c 		used = atomic_read(&pool->counter);
atomic_read       432 sound/core/seq/seq_memory.c 	while (atomic_read(&pool->counter) > 0)
atomic_read       498 sound/core/seq/seq_memory.c 	snd_iprintf(buffer, "%sCells in use       : %d\n", space, atomic_read(&pool->counter));
atomic_read        60 sound/core/seq/seq_memory.h 	return pool ? pool->total_elements - atomic_read(&pool->counter) : 0;
atomic_read       372 sound/drivers/dummy.c 	if (!atomic_read(&dpcm->running))
atomic_read       379 sound/drivers/dummy.c 	if (!atomic_read(&dpcm->running))
atomic_read        48 sound/drivers/pcsp/pcsp_input.c 	if (atomic_read(&pcsp_chip.timer_active) || !pcsp_chip.pcspkr)
atomic_read        31 sound/drivers/pcsp/pcsp_lib.c 	if (atomic_read(&pcsp_chip.timer_active)) {
atomic_read       135 sound/drivers/pcsp/pcsp_lib.c 	if (!atomic_read(&chip->timer_active) || !chip->playback_substream)
atomic_read       158 sound/drivers/pcsp/pcsp_lib.c 	if (atomic_read(&chip->timer_active)) {
atomic_read       179 sound/drivers/pcsp/pcsp_lib.c 	if (!atomic_read(&chip->timer_active))
atomic_read       318 sound/drivers/pcsp/pcsp_lib.c 	if (atomic_read(&chip->timer_active)) {
atomic_read       718 sound/isa/gus/gus_pcm.c 	if (!wait_event_timeout(pcmp->sleep, (atomic_read(&pcmp->dma_count) <= 0), 2*HZ))
atomic_read       357 sound/pci/echoaudio/echoaudio.c 	if (atomic_read(&chip->opencount) > 1 && chip->rate_set)
atomic_read       360 sound/pci/echoaudio/echoaudio.c 		chip->can_set_rate, atomic_read(&chip->opencount),
atomic_read       392 sound/pci/echoaudio/echoaudio.c 	if (atomic_read(&chip->opencount) > 1 && chip->rate_set)
atomic_read       395 sound/pci/echoaudio/echoaudio.c 		chip->can_set_rate, atomic_read(&chip->opencount),
atomic_read       433 sound/pci/echoaudio/echoaudio.c 	if (atomic_read(&chip->opencount) > 1 && chip->rate_set)
atomic_read       475 sound/pci/echoaudio/echoaudio.c 	if (atomic_read(&chip->opencount) > 1 && chip->rate_set)
atomic_read       498 sound/pci/echoaudio/echoaudio.c 	oc = atomic_read(&chip->opencount);
atomic_read      1447 sound/pci/echoaudio/echoaudio.c 		if (atomic_read(&chip->opencount)) {
atomic_read       305 sound/pci/echoaudio/mona_dsp.c 	if (atomic_read(&chip->opencount))
atomic_read       464 sound/pci/mixart/mixart.c 	while (atomic_read(&mgr->msg_processed) > 0) {
atomic_read       193 sound/pci/ymfpci/ymfpci_main.c 	if (atomic_read(&chip->interrupt_sleep_count)) {
atomic_read       816 sound/pci/ymfpci/ymfpci_main.c 		if (atomic_read(&chip->interrupt_sleep_count)) {
atomic_read      1625 sound/soc/codecs/hdac_hdmi.c 	if (atomic_read(&hdev->in_pm))
atomic_read        51 sound/soc/fsl/imx-pcm-fiq.c 	if (!atomic_read(&iprtd->playing) && !atomic_read(&iprtd->capturing))
atomic_read       132 sound/soc/fsl/imx-pcm-fiq.c 		if (!atomic_read(&iprtd->playing) &&
atomic_read       133 sound/soc/fsl/imx-pcm-fiq.c 				!atomic_read(&iprtd->capturing))
atomic_read        38 sound/soc/intel/atom/sst/sst_drv_interface.c #define GET_USAGE_COUNT(dev) (atomic_read(&dev->power.usage_count))
atomic_read       345 sound/usb/card.c 	if (!atomic_read(&chip->shutdown))
atomic_read       589 sound/usb/card.c 			if (atomic_read(&usb_chip[i]->shutdown)) {
atomic_read       717 sound/usb/card.c 			   !atomic_read(&chip->usage_count));
atomic_read       761 sound/usb/card.c 	if (atomic_read(&chip->shutdown)) {
atomic_read       788 sound/usb/card.c 	if (atomic_read(&chip->shutdown))
atomic_read       797 sound/usb/card.c 	if (atomic_read(&chip->shutdown))
atomic_read       369 sound/usb/endpoint.c 	if (unlikely(atomic_read(&ep->chip->shutdown)))
atomic_read       552 sound/usb/endpoint.c 	if (!force && atomic_read(&ep->chip->shutdown)) /* to be sure... */
atomic_read       926 sound/usb/endpoint.c 	if (atomic_read(&ep->chip->shutdown))
atomic_read       371 sound/usb/midi.c 				if (atomic_read(&in->urbs[j]->use_count))
atomic_read        71 sound/usb/pcm.c 	if (atomic_read(&subs->stream->chip->shutdown))
atomic_read        36 sound/usb/proc.c 	if (!atomic_read(&chip->shutdown))
atomic_read        43 sound/usb/proc.c 	if (!atomic_read(&chip->shutdown))
atomic_read       126 sound/usb/usx2y/us122l.c 	snd_printdd(KERN_DEBUG "%i\n", atomic_read(&us122l->mmap_count));
atomic_read       169 sound/usb/usx2y/us122l.c 	snd_printdd(KERN_DEBUG "%i\n", atomic_read(&us122l->mmap_count));
atomic_read       639 sound/usb/usx2y/us122l.c 	while (atomic_read(&us122l->mmap_count))
atomic_read       137 sound/usb/usx2y/usbusx2yaudio.c 	if (atomic_read(&subs->state) >= state_PRERUNNING)
atomic_read       205 sound/usb/usx2y/usbusx2yaudio.c 	state = atomic_read(&playbacksubs->state);
atomic_read       232 sound/usb/usx2y/usbusx2yaudio.c 	state = atomic_read(&capsubs->state);
atomic_read       254 sound/usb/usx2y/usbusx2yaudio.c 			snd_printdd("%i %p state=%i\n", s, subs, atomic_read(&subs->state));
atomic_read       261 sound/usb/usx2y/usbusx2yaudio.c 			if (atomic_read(&subs->state) >= state_PRERUNNING)
atomic_read       288 sound/usb/usx2y/usbusx2yaudio.c 	if (unlikely(atomic_read(&subs->state) < state_PREPARED)) {
atomic_read       306 sound/usb/usx2y/usbusx2yaudio.c 		    atomic_read(&capsubs->state) >= state_PREPARED &&
atomic_read       308 sound/usb/usx2y/usbusx2yaudio.c 		     atomic_read(&playbacksubs->state) < state_PREPARED)) {
atomic_read       462 sound/usb/usx2y/usbusx2yaudio.c 		if (subs != NULL && atomic_read(&subs->state) >= state_PREPARED)
atomic_read       495 sound/usb/usx2y/usbusx2yaudio.c 	if (atomic_read(&subs->state) != state_PREPARED)
atomic_read       524 sound/usb/usx2y/usbusx2yaudio.c 		if (atomic_read(&subs->state) == state_PREPARED &&
atomic_read       525 sound/usb/usx2y/usbusx2yaudio.c 		    atomic_read(&subs->usX2Y->subs[SNDRV_PCM_STREAM_CAPTURE]->state) >= state_PREPARED) {
atomic_read       534 sound/usb/usx2y/usbusx2yaudio.c 		if (atomic_read(&subs->state) >= state_PRERUNNING)
atomic_read       805 sound/usb/usx2y/usbusx2yaudio.c 		if (atomic_read(&playback_subs->state) < state_PREPARED) {
atomic_read       831 sound/usb/usx2y/usbusx2yaudio.c 	if (atomic_read(&capsubs->state) < state_PREPARED) {
atomic_read       843 sound/usb/usx2y/usbusx2yaudio.c 	if (subs != capsubs && atomic_read(&subs->state) < state_PREPARED)
atomic_read       125 sound/usb/usx2y/usx2yhwdeppcm.c 		if (atomic_read(&subs->state) != state_RUNNING)
atomic_read       168 sound/usb/usx2y/usx2yhwdeppcm.c 	state = atomic_read(&playbacksubs->state);
atomic_read       195 sound/usb/usx2y/usx2yhwdeppcm.c 	state = atomic_read(&capsubs->state);
atomic_read       224 sound/usb/usx2y/usx2yhwdeppcm.c 	if (unlikely(atomic_read(&subs->state) < state_PREPARED)) {
atomic_read       240 sound/usb/usx2y/usx2yhwdeppcm.c 	if (capsubs->completed_urb && atomic_read(&capsubs->state) >= state_PREPARED &&
atomic_read       242 sound/usb/usx2y/usx2yhwdeppcm.c 	    (playbacksubs->completed_urb || atomic_read(&playbacksubs->state) < state_PREPARED)) {
atomic_read       369 sound/usb/usx2y/usx2yhwdeppcm.c 		if (atomic_read(&playback_subs->state) < state_PREPARED) {
atomic_read       413 sound/usb/usx2y/usx2yhwdeppcm.c 		if (subs != NULL && atomic_read(&subs->state) >= state_PREPARED)
atomic_read       453 sound/usb/usx2y/usx2yhwdeppcm.c 	if (atomic_read(&subs->state) != state_PREPARED)
atomic_read       490 sound/usb/usx2y/usx2yhwdeppcm.c 	if (atomic_read(&capsubs->state) < state_PREPARED) {
atomic_read       505 sound/usb/usx2y/usx2yhwdeppcm.c 		if (atomic_read(&subs->state) < state_PREPARED) {
atomic_read       587 sound/xen/xen_snd_front_alsa.c 	new_hw_ptr = (snd_pcm_uframes_t)atomic_read(&stream->hw_ptr);
atomic_read       602 sound/xen/xen_snd_front_alsa.c 	return (snd_pcm_uframes_t)atomic_read(&stream->hw_ptr);
atomic_read        65 tools/include/linux/refcount.h 	return atomic_read(&r->refs);
atomic_read        78 tools/include/linux/refcount.h 	unsigned int old, new, val = atomic_read(&r->refs);
atomic_read       123 tools/include/linux/refcount.h 	unsigned int old, new, val = atomic_read(&r->refs);
atomic_read       906 virt/kvm/arm/arm.c 	int nrcpus = atomic_read(&kvm->online_vcpus);
atomic_read        58 virt/kvm/arm/vgic/vgic-debug.c 	int nr_cpus = atomic_read(&kvm->online_vcpus);
atomic_read       112 virt/kvm/arm/vgic/vgic-init.c 	if (atomic_read(&kvm->online_vcpus) > kvm->arch.max_vcpus) {
atomic_read       273 virt/kvm/arm/vgic/vgic-init.c 	if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
atomic_read      1216 virt/kvm/arm/vgic/vgic-its.c 	if (target_addr >= atomic_read(&kvm->online_vcpus))
atomic_read      1348 virt/kvm/arm/vgic/vgic-its.c 	if (target1_addr >= atomic_read(&kvm->online_vcpus) ||
atomic_read      1349 virt/kvm/arm/vgic/vgic-its.c 	    target2_addr >= atomic_read(&kvm->online_vcpus))
atomic_read      1836 virt/kvm/arm/vgic/vgic-its.c 	sz = atomic_read(&kvm->online_vcpus) * LPI_DEFAULT_PCPU_CACHE_SIZE;
atomic_read      2476 virt/kvm/arm/vgic/vgic-its.c 	    target_addr >= atomic_read(&kvm->online_vcpus))
atomic_read       289 virt/kvm/arm/vgic/vgic-kvm-device.c 	if (cpuid >= atomic_read(&dev->kvm->online_vcpus))
atomic_read       311 virt/kvm/arm/vgic/vgic-kvm-device.c 	unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
atomic_read        38 virt/kvm/arm/vgic/vgic-mmio-v2.c 		value |= (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5;
atomic_read       112 virt/kvm/arm/vgic/vgic-mmio-v2.c 	int nr_vcpus = atomic_read(&source_vcpu->kvm->online_vcpus);
atomic_read       174 virt/kvm/arm/vgic/vgic-mmio-v2.c 	u8 cpu_mask = GENMASK(atomic_read(&vcpu->kvm->online_vcpus) - 1, 0);
atomic_read       116 virt/kvm/arm/vgic/vgic-v4.c 	nr_vcpus = atomic_read(&kvm->online_vcpus);
atomic_read       289 virt/kvm/arm/vgic/vgic.h 		return atomic_read(&kvm->online_vcpus) * KVM_VGIC_V3_REDIST_SIZE;
atomic_read      2827 virt/kvm/kvm_main.c 	BUG_ON(kvm->vcpus[atomic_read(&kvm->online_vcpus)]);
atomic_read      2837 virt/kvm/kvm_main.c 	kvm->vcpus[atomic_read(&kvm->online_vcpus)] = vcpu;
atomic_read      3757 virt/kvm/kvm_main.c 		if (atomic_read(&hardware_enable_failed)) {