/linux-4.1.27/fs/fscache/ |
D | stats.c | 141 atomic_read(&fscache_n_cookie_index), in fscache_stats_show() 142 atomic_read(&fscache_n_cookie_data), in fscache_stats_show() 143 atomic_read(&fscache_n_cookie_special)); in fscache_stats_show() 146 atomic_read(&fscache_n_object_alloc), in fscache_stats_show() 147 atomic_read(&fscache_n_object_no_alloc), in fscache_stats_show() 148 atomic_read(&fscache_n_object_avail), in fscache_stats_show() 149 atomic_read(&fscache_n_object_dead)); in fscache_stats_show() 151 atomic_read(&fscache_n_checkaux_none), in fscache_stats_show() 152 atomic_read(&fscache_n_checkaux_okay), in fscache_stats_show() 153 atomic_read(&fscache_n_checkaux_update), in fscache_stats_show() [all …]
|
D | histogram.c | 41 n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); in fscache_histogram_show() 42 n[1] = atomic_read(&fscache_ops_histogram[index]); in fscache_histogram_show() 43 n[2] = atomic_read(&fscache_objs_histogram[index]); in fscache_histogram_show() 44 n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); in fscache_histogram_show() 45 n[4] = atomic_read(&fscache_retrieval_histogram[index]); in fscache_histogram_show()
|
D | operation.c | 34 op->object->debug_id, op->debug_id, atomic_read(&op->usage)); in fscache_enqueue_operation() 39 ASSERTCMP(atomic_read(&op->usage), >, 0); in fscache_enqueue_operation() 91 ASSERTCMP(atomic_read(&op->usage), >, 0); in fscache_submit_exclusive_op() 193 object->debug_id, op->debug_id, atomic_read(&op->usage)); in fscache_submit_op() 196 ASSERTCMP(atomic_read(&op->usage), >, 0); in fscache_submit_op() 305 ASSERTCMP(atomic_read(&op->usage), >, 0); in fscache_cancel_op() 404 op->object->debug_id, op->debug_id, atomic_read(&op->usage)); in fscache_put_operation() 406 ASSERTCMP(atomic_read(&op->usage), >, 0); in fscache_put_operation() 490 ASSERTCMP(atomic_read(&op->usage), ==, 0); in fscache_operation_gc() 520 op->object->debug_id, op->debug_id, atomic_read(&op->usage)); in fscache_op_work_func()
|
D | cookie.c | 518 ASSERTCMP(atomic_read(&cookie->n_active), >, 0); in __fscache_disable_cookie() 520 if (atomic_read(&cookie->n_children) != 0) { in __fscache_disable_cookie() 595 atomic_read(&cookie->n_active), retire); in __fscache_relinquish_cookie() 608 ASSERTCMP(atomic_read(&cookie->parent->usage), >, 0); in __fscache_relinquish_cookie() 609 ASSERTCMP(atomic_read(&cookie->parent->n_children), >, 0); in __fscache_relinquish_cookie() 614 ASSERTCMP(atomic_read(&cookie->usage), >, 0); in __fscache_relinquish_cookie() 640 BUG_ON(atomic_read(&cookie->usage) <= 0); in __fscache_cookie_put()
|
/linux-4.1.27/fs/btrfs/ |
D | locking.c | 45 if (atomic_read(&eb->blocking_writers) == 0) { in btrfs_set_lock_blocking_rw() 46 WARN_ON(atomic_read(&eb->spinning_writers) != 1); in btrfs_set_lock_blocking_rw() 55 WARN_ON(atomic_read(&eb->spinning_readers) == 0); in btrfs_set_lock_blocking_rw() 78 BUG_ON(atomic_read(&eb->blocking_writers) != 1); in btrfs_clear_lock_blocking_rw() 80 WARN_ON(atomic_read(&eb->spinning_writers)); in btrfs_clear_lock_blocking_rw() 86 BUG_ON(atomic_read(&eb->blocking_readers) == 0); in btrfs_clear_lock_blocking_rw() 103 BUG_ON(!atomic_read(&eb->blocking_writers) && in btrfs_tree_read_lock() 107 if (atomic_read(&eb->blocking_writers) && in btrfs_tree_read_lock() 120 if (atomic_read(&eb->blocking_writers)) { in btrfs_tree_read_lock() 123 atomic_read(&eb->blocking_writers) == 0); in btrfs_tree_read_lock() [all …]
|
D | reada.c | 766 if (atomic_read(&device->reada_in_flight) < in __reada_start_machine() 818 atomic_read(&device->reada_in_flight)); in dump_devs() 959 while (atomic_read(&rc->elems)) { in btrfs_reada_wait() 960 wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0, in btrfs_reada_wait() 963 atomic_read(&rc->elems) < 10 ? 1 : 0); in btrfs_reada_wait() 966 dump_devs(rc->root->fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0); in btrfs_reada_wait() 977 while (atomic_read(&rc->elems)) { in btrfs_reada_wait() 978 wait_event(rc->wait, atomic_read(&rc->elems) == 0); in btrfs_reada_wait()
|
/linux-4.1.27/net/netfilter/ipvs/ |
D | ip_vs_nq.c | 50 return atomic_read(&dest->activeconns) + 1; in ip_vs_nq_dest_overhead() 82 !atomic_read(&dest->weight)) in ip_vs_nq_schedule() 88 if (atomic_read(&dest->activeconns) == 0) { in ip_vs_nq_schedule() 95 ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_nq_schedule() 96 (__s64)doh * atomic_read(&least->weight))) { in ip_vs_nq_schedule() 112 atomic_read(&least->activeconns), in ip_vs_nq_schedule() 113 atomic_read(&least->refcnt), in ip_vs_nq_schedule() 114 atomic_read(&least->weight), loh); in ip_vs_nq_schedule()
|
D | ip_vs_sed.c | 54 return atomic_read(&dest->activeconns) + 1; in ip_vs_sed_dest_overhead() 85 atomic_read(&dest->weight) > 0) { in ip_vs_sed_schedule() 102 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_sed_schedule() 103 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_sed_schedule() 113 atomic_read(&least->activeconns), in ip_vs_sed_schedule() 114 atomic_read(&least->refcnt), in ip_vs_sed_schedule() 115 atomic_read(&least->weight), loh); in ip_vs_sed_schedule()
|
D | ip_vs_lblcr.c | 177 if ((atomic_read(&least->weight) > 0) in ip_vs_dest_set_min() 193 if (((__s64)loh * atomic_read(&dest->weight) > in ip_vs_dest_set_min() 194 (__s64)doh * atomic_read(&least->weight)) in ip_vs_dest_set_min() 206 atomic_read(&least->activeconns), in ip_vs_dest_set_min() 207 atomic_read(&least->refcnt), in ip_vs_dest_set_min() 208 atomic_read(&least->weight), loh); in ip_vs_dest_set_min() 226 if (atomic_read(&most->weight) > 0) { in ip_vs_dest_set_max() 239 if (((__s64)moh * atomic_read(&dest->weight) < in ip_vs_dest_set_max() 240 (__s64)doh * atomic_read(&most->weight)) in ip_vs_dest_set_max() 241 && (atomic_read(&dest->weight) > 0)) { in ip_vs_dest_set_max() [all …]
|
D | ip_vs_wlc.c | 57 atomic_read(&dest->weight) > 0) { in ip_vs_wlc_schedule() 74 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_wlc_schedule() 75 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_wlc_schedule() 85 atomic_read(&least->activeconns), in ip_vs_wlc_schedule() 86 atomic_read(&least->refcnt), in ip_vs_wlc_schedule() 87 atomic_read(&least->weight), loh); in ip_vs_wlc_schedule()
|
D | ip_vs_fo.c | 39 atomic_read(&dest->weight) > hw) { in ip_vs_fo_schedule() 41 hw = atomic_read(&dest->weight); in ip_vs_fo_schedule() 49 atomic_read(&hweight->activeconns), in ip_vs_fo_schedule() 50 atomic_read(&hweight->weight)); in ip_vs_fo_schedule()
|
D | ip_vs_lblc.c | 315 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblc_check_expire() 320 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblc_check_expire() 423 if (atomic_read(&dest->weight) > 0) { in __ip_vs_lblc_schedule() 440 if ((__s64)loh * atomic_read(&dest->weight) > in __ip_vs_lblc_schedule() 441 (__s64)doh * atomic_read(&least->weight)) { in __ip_vs_lblc_schedule() 451 atomic_read(&least->activeconns), in __ip_vs_lblc_schedule() 452 atomic_read(&least->refcnt), in __ip_vs_lblc_schedule() 453 atomic_read(&least->weight), loh); in __ip_vs_lblc_schedule() 466 if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) { in is_overloaded() 470 if (atomic_read(&d->activeconns)*2 in is_overloaded() [all …]
|
D | ip_vs_wrr.c | 81 weight = atomic_read(&dest->weight); in ip_vs_wrr_gcd_weight() 102 new_weight = atomic_read(&dest->weight); in ip_vs_wrr_max_weight() 186 atomic_read(&dest->weight) >= mark->cw) in ip_vs_wrr_schedule() 220 atomic_read(&dest->activeconns), in ip_vs_wrr_schedule() 221 atomic_read(&dest->refcnt), in ip_vs_wrr_schedule() 222 atomic_read(&dest->weight)); in ip_vs_wrr_schedule()
|
D | ip_vs_rr.c | 76 atomic_read(&dest->weight) > 0) in ip_vs_rr_schedule() 99 atomic_read(&dest->activeconns), in ip_vs_rr_schedule() 100 atomic_read(&dest->refcnt), atomic_read(&dest->weight)); in ip_vs_rr_schedule()
|
D | ip_vs_lc.c | 48 atomic_read(&dest->weight) == 0) in ip_vs_lc_schedule() 64 atomic_read(&least->activeconns), in ip_vs_lc_schedule() 65 atomic_read(&least->inactconns)); in ip_vs_lc_schedule()
|
D | ip_vs_conn.c | 250 ret = atomic_read(&cp->refcnt) ? false : true; in ip_vs_conn_unlink() 301 if (!cp && atomic_read(&ip_vs_conn_no_cport_cnt)) { in ip_vs_conn_in_get() 554 return atomic_read(&dest->activeconns) in ip_vs_dest_totalconns() 555 + atomic_read(&dest->inactconns); in ip_vs_dest_totalconns() 575 conn_flags = atomic_read(&dest->conn_flags); in ip_vs_bind_dest() 601 cp->flags, atomic_read(&cp->refcnt), in ip_vs_bind_dest() 602 atomic_read(&dest->refcnt)); in ip_vs_bind_dest() 672 if (pd && atomic_read(&pd->appcnt)) in ip_vs_try_bind_dest() 698 cp->flags, atomic_read(&cp->refcnt), in ip_vs_unbind_dest() 699 atomic_read(&dest->refcnt)); in ip_vs_unbind_dest() [all …]
|
D | ip_vs_sh.c | 81 return atomic_read(&dest->weight) <= 0 || in is_unavailable() 196 atomic_read(&dest->weight)); in ip_vs_sh_reassign() 199 if (++d_count >= atomic_read(&dest->weight)) { in ip_vs_sh_reassign()
|
/linux-4.1.27/drivers/staging/lustre/include/linux/libcfs/ |
D | libcfs_private.h | 110 atomic_read(&libcfs_kmemory) 253 LASSERTF(atomic_read(a) == v, \ 254 "value: %d\n", atomic_read((a))); \ 260 LASSERTF(atomic_read(a) != v, \ 261 "value: %d\n", atomic_read((a))); \ 267 LASSERTF(atomic_read(a) < v, \ 268 "value: %d\n", atomic_read((a))); \ 274 LASSERTF(atomic_read(a) <= v, \ 275 "value: %d\n", atomic_read((a))); \ 281 LASSERTF(atomic_read(a) > v, \ [all …]
|
/linux-4.1.27/arch/mips/kernel/ |
D | sync-r4k.c | 61 while (atomic_read(&count_count_start) != 1) in synchronise_count_master() 78 while (atomic_read(&count_count_stop) != 1) in synchronise_count_master() 108 while (atomic_read(&count_start_flag) != cpu) in synchronise_count_slave() 112 initcount = atomic_read(&count_reference); in synchronise_count_slave() 116 while (atomic_read(&count_count_start) != 2) in synchronise_count_slave() 126 while (atomic_read(&count_count_stop) != 2) in synchronise_count_slave()
|
D | spinlock_test.c | 66 while (atomic_read(&s->enter_wait)) in multi_other() 73 while (atomic_read(&s->start_wait)) in multi_other() 85 while (atomic_read(&s->exit_wait)) in multi_other()
|
/linux-4.1.27/sound/core/seq/ |
D | seq_lock.c | 33 if (atomic_read(lockp) < 0) { in snd_use_lock_sync_helper() 34 pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper() 37 while (atomic_read(lockp) > 0) { in snd_use_lock_sync_helper() 39 pr_warn("ALSA: seq_lock: timeout [%d left] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper()
|
D | seq_memory.c | 37 return pool->total_elements - atomic_read(&pool->counter); in snd_seq_pool_available() 267 used = atomic_read(&pool->counter); in snd_seq_cell_alloc() 435 while (atomic_read(&pool->counter) > 0) { in snd_seq_pool_done() 437 pr_warn("ALSA: snd_seq_pool_done timeout: %d cells remain\n", atomic_read(&pool->counter)); in snd_seq_pool_done() 518 snd_iprintf(buffer, "%sCells in use : %d\n", space, atomic_read(&pool->counter)); in snd_seq_info_pool()
|
/linux-4.1.27/drivers/staging/lustre/lustre/osc/ |
D | osc_page.c | 401 atomic_read(&obj->oo_nr_reads), in osc_page_print() 403 atomic_read(&obj->oo_nr_writes), in osc_page_print() 596 int pages = atomic_read(&cli->cl_lru_in_list) >> 1; in osc_cache_too_much() 598 if (atomic_read(&osc_lru_waiters) > 0 && in osc_cache_too_much() 599 atomic_read(cli->cl_lru_left) < lru_shrink_max) in osc_cache_too_much() 605 if (atomic_read(cli->cl_lru_left) < cache->ccc_lru_max >> 4) { in osc_cache_too_much() 608 tmp = cache->ccc_lru_max / atomic_read(&cache->ccc_users); in osc_cache_too_much() 661 LASSERT(atomic_read(&cli->cl_lru_in_list) >= 0); in osc_lru_shrink() 662 if (atomic_read(&cli->cl_lru_in_list) == 0 || target <= 0) in osc_lru_shrink() 674 maxscan = min(target << 1, atomic_read(&cli->cl_lru_in_list)); in osc_lru_shrink() [all …]
|
D | lproc_osc.c | 171 (atomic_read(&cli->cl_lru_in_list) + in osc_cached_mb_seq_show() 172 atomic_read(&cli->cl_lru_busy)) >> shift, in osc_cached_mb_seq_show() 173 atomic_read(&cli->cl_lru_busy)); in osc_cached_mb_seq_show() 205 rc = atomic_read(&cli->cl_lru_in_list) - pages_number; in osc_cached_mb_seq_write() 407 seq_printf(m, "%u\n", atomic_read(&obd->u.cli.cl_resends)); in osc_resend_count_seq_show() 477 seq_printf(m, "%u\n", atomic_read(&obd->u.cli.cl_destroy_in_flight)); in osc_destroys_in_flight_seq_show() 603 atomic_read(&cli->cl_pending_w_pages)); in osc_rpc_stats_seq_show() 605 atomic_read(&cli->cl_pending_r_pages)); in osc_rpc_stats_seq_show()
|
D | osc_cache.c | 115 atomic_read(&__ext->oe_refc), \ 116 atomic_read(&__ext->oe_users), \ 191 if (atomic_read(&ext->oe_refc) <= 0) { in osc_extent_sanity_check0() 196 if (atomic_read(&ext->oe_refc) < atomic_read(&ext->oe_users)) { in osc_extent_sanity_check0() 209 if (atomic_read(&ext->oe_users) == 0) { in osc_extent_sanity_check0() 232 if (atomic_read(&ext->oe_users) > 0) { in osc_extent_sanity_check0() 373 LASSERT(atomic_read(&ext->oe_refc) >= 0); in osc_extent_get() 380 LASSERT(atomic_read(&ext->oe_refc) > 0); in osc_extent_put() 383 LASSERT(atomic_read(&ext->oe_users) == 0); in osc_extent_put() 402 LASSERT(atomic_read(&ext->oe_refc) > 1); in osc_extent_put_trust() [all …]
|
/linux-4.1.27/net/batman-adv/ |
D | gateway_common.c | 121 gw_mode = atomic_read(&bat_priv->gw_mode); in batadv_gw_tvlv_container_update() 129 down = atomic_read(&bat_priv->gw.bandwidth_down); in batadv_gw_tvlv_container_update() 130 up = atomic_read(&bat_priv->gw.bandwidth_up); in batadv_gw_tvlv_container_update() 146 down_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_down); in batadv_gw_bandwidth_set() 147 up_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_up); in batadv_gw_bandwidth_set() 217 (atomic_read(&bat_priv->gw_mode) == BATADV_GW_MODE_CLIENT) && in batadv_gw_tvlv_ogm_handler_v1() 218 (atomic_read(&bat_priv->gw_sel_class) > 2)) in batadv_gw_tvlv_ogm_handler_v1()
|
D | sysfs.c | 141 atomic_read(&bat_priv->_name) == 0 ? \ 173 return sprintf(buff, "%i\n", atomic_read(&bat_priv->_name)); \ 209 atomic_read(&vlan->_name) == 0 ? \ 248 if (atomic_read(attr) == enabled) in batadv_store_bool_attr() 252 atomic_read(attr) == 1 ? "enabled" : "disabled", in batadv_store_bool_attr() 303 if (atomic_read(attr) == uint_val) in batadv_store_uint_attr() 307 attr_name, atomic_read(attr), uint_val); in batadv_store_uint_attr() 351 switch (atomic_read(&bat_priv->gw_mode)) { in batadv_show_gw_mode() 400 if (atomic_read(&bat_priv->gw_mode) == gw_mode_tmp) in batadv_store_gw_mode() 403 switch (atomic_read(&bat_priv->gw_mode)) { in batadv_store_gw_mode() [all …]
|
D | gateway_client.c | 165 switch (atomic_read(&bat_priv->gw_sel_class)) { in batadv_gw_get_best_gw_node() 228 if (atomic_read(&bat_priv->gw_mode) != BATADV_GW_MODE_CLIENT) in batadv_gw_check_client_stop() 255 if (atomic_read(&bat_priv->gw_mode) != BATADV_GW_MODE_CLIENT) in batadv_gw_election() 373 if ((atomic_read(&bat_priv->gw_sel_class) > 3) && in batadv_gw_check_election() 374 (orig_tq_avg - gw_tq_avg < atomic_read(&bat_priv->gw_sel_class))) in batadv_gw_check_election() 555 atomic_read(&bat_priv->mesh_state) == BATADV_MESH_ACTIVE) in batadv_gw_node_purge() 818 switch (atomic_read(&bat_priv->gw_mode)) { in batadv_gw_out_of_range()
|
D | bridge_loop_avoidance.c | 520 if (!atomic_read(&backbone_gw->request_sent)) { in batadv_bla_send_request() 680 if (atomic_read(&backbone_gw->request_sent)) { in batadv_handle_announce() 1020 if (atomic_read(&backbone_gw->request_sent)) in batadv_bla_purge_backbone_gw() 1104 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) in batadv_bla_update_orig_address() 1163 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) in batadv_bla_periodic_work() 1192 if (atomic_read(&backbone_gw->request_sent) == 0) in batadv_bla_periodic_work() 1361 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) in batadv_bla_is_backbone_gw_orig() 1400 if (!atomic_read(&orig_node->bat_priv->bridge_loop_avoidance)) in batadv_bla_is_backbone_gw() 1470 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) in batadv_bla_rx() 1473 if (unlikely(atomic_read(&bat_priv->bla.num_requests))) in batadv_bla_rx() [all …]
|
D | multicast.c | 346 if (!atomic_read(&bat_priv->multicast_mode)) in batadv_mcast_forw_mode_check() 349 if (atomic_read(&bat_priv->mcast.num_disabled)) in batadv_mcast_forw_mode_check() 378 return atomic_read(&bat_priv->mcast.num_want_all_ipv4); in batadv_mcast_forw_want_all_ip_count() 380 return atomic_read(&bat_priv->mcast.num_want_all_ipv6); in batadv_mcast_forw_want_all_ip_count() 538 atomic_read(&bat_priv->mcast.num_want_all_unsnoopables); in batadv_mcast_forw_mode()
|
D | translation-table.c | 216 count = atomic_read(&tt_global_entry->orig_list_count); in batadv_tt_global_hash_count() 451 tt_local_entries += atomic_read(&vlan->tt.num_entries); in batadv_tt_local_table_transmit_size() 563 packet_size_max = atomic_read(&bat_priv->packet_size_max); in batadv_tt_local_add() 584 (uint8_t)atomic_read(&bat_priv->tt.vn)); in batadv_tt_local_add() 717 num_entries += atomic_read(&vlan->tt.num_entries); in batadv_tt_prepare_tvlv_global_data() 737 (*tt_data)->ttvn = atomic_read(&orig_node->last_ttvn); in batadv_tt_prepare_tvlv_global_data() 789 num_entries += atomic_read(&vlan->tt.num_entries); in batadv_tt_prepare_tvlv_local_data() 809 (*tt_data)->ttvn = atomic_read(&bat_priv->tt.vn); in batadv_tt_prepare_tvlv_local_data() 842 tt_diff_entries_num = atomic_read(&bat_priv->tt.local_changes); in batadv_tt_tvlv_container_update() 927 net_dev->name, (uint8_t)atomic_read(&bat_priv->tt.vn)); in batadv_tt_local_seq_print_text() [all …]
|
/linux-4.1.27/kernel/ |
D | cred.c | 72 return atomic_read(&cred->subscribers); in read_cred_subscribers() 98 atomic_read(&cred->usage) != 0 || in put_cred_rcu() 103 atomic_read(&cred->usage), in put_cred_rcu() 106 if (atomic_read(&cred->usage) != 0) in put_cred_rcu() 108 cred, atomic_read(&cred->usage)); in put_cred_rcu() 132 atomic_read(&cred->usage), in __put_cred() 135 BUG_ON(atomic_read(&cred->usage) != 0); in __put_cred() 156 atomic_read(&tsk->cred->usage), in exit_creds() 332 p->cred, atomic_read(&p->cred->usage), in copy_creds() 423 atomic_read(&new->usage), in commit_creds() [all …]
|
D | smpboot.c | 327 return atomic_read(&per_cpu(cpu_hotplug_state, cpu)); in cpu_report_state() 349 switch (atomic_read(&per_cpu(cpu_hotplug_state, cpu))) { in cpu_check_up_prepare() 418 if (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) == CPU_DEAD) in cpu_wait_death() 423 while (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) != CPU_DEAD) { in cpu_wait_death() 431 oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu)); in cpu_wait_death() 462 oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu)); in cpu_report_death()
|
/linux-4.1.27/drivers/s390/scsi/ |
D | zfcp_erp.c | 95 if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_ERP_INUSE) in zfcp_erp_action_dismiss_lun() 103 if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_INUSE) in zfcp_erp_action_dismiss_port() 118 if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_INUSE) in zfcp_erp_action_dismiss_adapter() 139 l_status = atomic_read(&zfcp_sdev->status); in zfcp_erp_required_act() 142 p_status = atomic_read(&port->status); in zfcp_erp_required_act() 150 p_status = atomic_read(&port->status); in zfcp_erp_required_act() 155 p_status = atomic_read(&port->status); in zfcp_erp_required_act() 158 a_status = atomic_read(&adapter->status); in zfcp_erp_required_act() 168 a_status = atomic_read(&adapter->status); in zfcp_erp_required_act() 199 if (!(atomic_read(&zfcp_sdev->status) & in zfcp_erp_setup_act() [all …]
|
D | zfcp_sysfs.c | 59 ZFCP_DEFINE_A_ATTR(status, "0x%08x\n", atomic_read(&adapter->status)); 68 ZFCP_DEFINE_A_ATTR(in_recovery, "%d\n", (atomic_read(&adapter->status) & 72 atomic_read(&port->status)); 74 (atomic_read(&port->status) & 95 if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED) in zfcp_sysfs_port_failed_show() 131 status = atomic_read(&sdev_to_zfcp(sdev)->status); in zfcp_sysfs_unit_failed_show() 176 if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_FAILED) in zfcp_sysfs_adapter_failed_show() 263 if (atomic_read(&port->units) > 0) { in zfcp_sysfs_port_remove_store() 465 (atomic_read(&zfcp_sdev->status) & 473 unsigned int status = atomic_read(&sdev_to_zfcp(sdev)->status); in zfcp_sysfs_scsi_zfcp_failed_show() [all …]
|
D | zfcp_qdio.c | 56 used = QDIO_MAX_BUFFERS_PER_Q - atomic_read(&qdio->req_q_free); in zfcp_qdio_account() 210 if (atomic_read(&qdio->req_q_free) || in zfcp_qdio_sbal_check() 211 !(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) in zfcp_qdio_sbal_check() 233 if (!(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) in zfcp_qdio_sbal_get() 347 if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) in zfcp_qdio_close() 360 count = atomic_read(&qdio->req_q_free); in zfcp_qdio_close() 384 if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP) in zfcp_qdio_open() 497 if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_SIOSL_ISSUED) in zfcp_qdio_siosl()
|
D | zfcp_dbf.c | 235 rec->adapter_status = atomic_read(&adapter->status); in zfcp_dbf_set_common() 237 rec->port_status = atomic_read(&port->status); in zfcp_dbf_set_common() 242 rec->lun_status = atomic_read(&sdev_to_zfcp(sdev)->status); in zfcp_dbf_set_common() 313 atomic_read(&sdev_to_zfcp(erp->sdev)->erp_counter); in zfcp_dbf_rec_run() 315 rec->u.run.rec_count = atomic_read(&erp->port->erp_counter); in zfcp_dbf_rec_run() 317 rec->u.run.rec_count = atomic_read(&erp->adapter->erp_counter); in zfcp_dbf_rec_run()
|
D | zfcp_scsi.c | 80 status = atomic_read(&zfcp_sdev->status); in zfcp_scsi_queuecommand() 82 !(atomic_read(&zfcp_sdev->port->status) & in zfcp_scsi_queuecommand() 186 if (!(atomic_read(&adapter->status) & in zfcp_scsi_eh_abort_handler() 230 if (!(atomic_read(&adapter->status) & in zfcp_task_mgmt_function() 504 int status = atomic_read(&adapter->status); in zfcp_get_host_port_state() 646 data_div = atomic_read(&adapter->status) & in zfcp_scsi_set_prot()
|
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
D | connection.c | 84 conn, atomic_read(&conn->c_refcount), in ptlrpc_connection_get() 97 LASSERT(atomic_read(&conn->c_refcount) > 1); in ptlrpc_connection_put() 119 conn, atomic_read(&conn->c_refcount), in ptlrpc_connection_put() 131 conn, atomic_read(&conn->c_refcount), in ptlrpc_connection_addref() 227 LASSERTF(atomic_read(&conn->c_refcount) == 0, in conn_exit() 229 atomic_read(&conn->c_refcount)); in conn_exit()
|
D | ptlrpcd.c | 168 i = atomic_read(&set->set_remaining); in ptlrpcd_add_rqset() 203 rc = atomic_read(&src->set_new_count); in ptlrpcd_steal_rqset() 271 if (atomic_read(&set->set_new_count)) { in ptlrpcd_check() 276 atomic_add(atomic_read(&set->set_new_count), in ptlrpcd_check() 306 if (atomic_read(&set->set_remaining)) in ptlrpcd_check() 325 rc = atomic_read(&set->set_new_count); in ptlrpcd_check() 351 if (atomic_read(&ps->set_new_count)) { in ptlrpcd_check()
|
D | import.c | 321 (atomic_read(&imp->imp_inflight) == 0), in ptlrpc_invalidate_import() 328 atomic_read(&imp->imp_inflight)); in ptlrpc_invalidate_import() 331 if (atomic_read(&imp->imp_inflight) == 0) { in ptlrpc_invalidate_import() 332 int count = atomic_read(&imp->imp_unregistering); in ptlrpc_invalidate_import() 368 atomic_read(&imp-> in ptlrpc_invalidate_import() 379 LASSERT(atomic_read(&imp->imp_inflight) == 0); in ptlrpc_invalidate_import() 461 if (atomic_read(&imp->imp_inval_count) > 0) { in ptlrpc_reconnect_import() 465 (atomic_read(&imp->imp_inval_count) == 0), in ptlrpc_reconnect_import() 469 atomic_read(&imp->imp_inval_count)); in ptlrpc_reconnect_import() 1251 LASSERT(atomic_read(&imp->imp_replay_inflight) == 0); in signal_completed_replay() [all …]
|
/linux-4.1.27/fs/afs/ |
D | server.c | 94 _leave(" = %p{%d}", server, atomic_read(&server->usage)); in afs_alloc_server() 143 _leave(" = %p{%d}", server, atomic_read(&server->usage)); in afs_lookup_server() 157 _leave(" = %p{%d}", server, atomic_read(&server->usage)); in afs_lookup_server() 224 _enter("%p{%d}", server, atomic_read(&server->usage)); in afs_put_server() 226 _debug("PUT SERVER %d", atomic_read(&server->usage)); in afs_put_server() 228 ASSERTCMP(atomic_read(&server->usage), >, 0); in afs_put_server() 238 if (atomic_read(&server->usage) == 0) { in afs_put_server() 261 ASSERTCMP(atomic_read(&server->cb_break_n), ==, 0); in afs_destroy_server() 294 if (atomic_read(&server->usage) > 0) { in afs_reap_server()
|
D | cell.c | 348 _enter("%p{%d,%s}", cell, atomic_read(&cell->usage), cell->name); in afs_put_cell() 350 ASSERTCMP(atomic_read(&cell->usage), >, 0); in afs_put_cell() 379 _enter("%p{%d,%s}", cell, atomic_read(&cell->usage), cell->name); in afs_cell_destroy() 381 ASSERTCMP(atomic_read(&cell->usage), >=, 0); in afs_cell_destroy() 385 if (atomic_read(&cell->usage) > 0) { in afs_cell_destroy() 392 while (atomic_read(&cell->usage) > 0) { in afs_cell_destroy() 402 ASSERTCMP(atomic_read(&cell->usage), ==, 0); in afs_cell_destroy() 451 cell->name, atomic_read(&cell->usage)); in afs_cell_purge()
|
D | rxrpc.c | 134 ASSERTCMP(atomic_read(&afs_outstanding_skbs), ==, 0); in afs_close_socket() 135 ASSERTCMP(atomic_read(&afs_outstanding_calls), ==, 0); in afs_close_socket() 146 _debug("DLVR NULL [%d]", atomic_read(&afs_outstanding_skbs)); in afs_data_delivered() 150 skb, skb->mark, atomic_read(&afs_outstanding_skbs)); in afs_data_delivered() 163 _debug("FREE NULL [%d]", atomic_read(&afs_outstanding_skbs)); in afs_free_skb() 167 skb, skb->mark, atomic_read(&afs_outstanding_skbs)); in afs_free_skb() 180 call, call->type->name, atomic_read(&afs_outstanding_calls)); in afs_free_call() 228 call, type->name, atomic_read(&afs_outstanding_calls)); in afs_alloc_flat_call() 355 atomic_read(&afs_outstanding_calls)); in afs_make_call() 435 skb, skb->mark, atomic_read(&afs_outstanding_skbs)); in afs_rx_interceptor() [all …]
|
/linux-4.1.27/kernel/locking/ |
D | qrwlock.c | 76 while (atomic_read(&lock->cnts) & _QW_WMASK) in queue_read_lock_slowpath() 101 if (!atomic_read(&lock->cnts) && in queue_write_lock_slowpath() 110 cnts = atomic_read(&lock->cnts); in queue_write_lock_slowpath() 121 cnts = atomic_read(&lock->cnts); in queue_write_lock_slowpath()
|
D | percpu-rwsem.c | 60 if (likely(!atomic_read(&brw->write_ctr))) { in update_fast_ctr() 151 wait_event(brw->write_waitq, !atomic_read(&brw->slow_read_ctr)); in percpu_down_write()
|
/linux-4.1.27/net/rxrpc/ |
D | ar-peer.c | 123 atomic_read(&peer->usage), in rxrpc_get_peer() 127 if (atomic_read(&peer->usage) > 0 && in rxrpc_get_peer() 148 if (atomic_read(&peer->usage) > 0 && in rxrpc_get_peer() 160 usage = atomic_read(&peer->usage); in rxrpc_get_peer() 209 if (atomic_read(&peer->usage) > 0 && in rxrpc_find_peer() 243 _enter("%p{u=%d}", peer, atomic_read(&peer->usage)); in rxrpc_put_peer() 245 ASSERTCMP(atomic_read(&peer->usage), >, 0); in rxrpc_put_peer() 264 _enter("%p{%d}", peer, atomic_read(&peer->usage)); in rxrpc_destroy_peer()
|
D | ar-transport.c | 118 usage = atomic_read(&trans->usage); in rxrpc_get_transport() 188 _enter("%p{u=%d}", trans, atomic_read(&trans->usage)); in rxrpc_put_transport() 190 ASSERTCMP(atomic_read(&trans->usage), >, 0); in rxrpc_put_transport() 236 trans->debug_id, atomic_read(&trans->usage), in rxrpc_transport_reaper() 239 if (likely(atomic_read(&trans->usage) > 0)) in rxrpc_transport_reaper() 263 ASSERTCMP(atomic_read(&trans->usage), ==, 0); in rxrpc_transport_reaper()
|
D | ar-proc.c | 81 atomic_read(&call->usage), in rxrpc_call_seq_show() 164 atomic_read(&conn->usage), in rxrpc_connection_seq_show() 167 atomic_read(&conn->serial), in rxrpc_connection_seq_show() 168 atomic_read(&conn->hi_serial)); in rxrpc_connection_seq_show()
|
D | ar-connection.c | 155 _leave(" = %p [extant %d]", bundle, atomic_read(&bundle->usage)); in rxrpc_get_bundle() 168 _leave(" = %p [second %d]", bundle, atomic_read(&bundle->usage)); in rxrpc_get_bundle() 178 _enter("%p,%p{%d}",trans, bundle, atomic_read(&bundle->usage)); in rxrpc_put_bundle() 719 _leave(" = %p {u=%d}", conn, atomic_read(&conn->usage)); in rxrpc_incoming_connection() 807 conn, atomic_read(&conn->usage), conn->debug_id); in rxrpc_put_connection() 809 ASSERTCMP(atomic_read(&conn->usage), >, 0); in rxrpc_put_connection() 825 _enter("%p{%d}", conn, atomic_read(&conn->usage)); in rxrpc_destroy_connection() 827 ASSERTCMP(atomic_read(&conn->usage), ==, 0); in rxrpc_destroy_connection() 861 conn->debug_id, atomic_read(&conn->usage), in rxrpc_connection_reaper() 864 if (likely(atomic_read(&conn->usage) > 0)) in rxrpc_connection_reaper() [all …]
|
D | ar-call.c | 399 _leave(" = %p [extant %d]", call, atomic_read(&call->usage)); in rxrpc_get_client_call() 407 _leave(" = %p [second %d]", call, atomic_read(&call->usage)); in rxrpc_get_client_call() 606 _leave(" = %p [%d]", call, atomic_read(&call->usage)); in rxrpc_find_server_call() 619 call->debug_id, atomic_read(&call->usage), in rxrpc_release_call() 620 atomic_read(&call->ackr_not_idle), in rxrpc_release_call() 815 _enter("%p{u=%d}", call, atomic_read(&call->usage)); in __rxrpc_put_call() 817 ASSERTCMP(atomic_read(&call->usage), >, 0); in __rxrpc_put_call() 903 call, atomic_read(&call->usage), call->channel, call->conn); in rxrpc_destroy_call() 932 switch (atomic_read(&call->usage)) { in rxrpc_destroy_all_calls() 945 call, atomic_read(&call->usage), in rxrpc_destroy_all_calls() [all …]
|
D | ar-local.c | 231 _enter("%p{u=%d}", local, atomic_read(&local->usage)); in rxrpc_put_local() 233 ASSERTCMP(atomic_read(&local->usage), >, 0); in rxrpc_put_local() 254 _enter("%p{%d}", local, atomic_read(&local->usage)); in rxrpc_destroy_local() 259 if (atomic_read(&local->usage) > 0) { in rxrpc_destroy_local()
|
/linux-4.1.27/arch/x86/platform/uv/ |
D | uv_nmi.c | 221 nmi = atomic_read(&hub_nmi->in_nmi); in uv_check_nmi() 243 nmi = atomic_read(&hub_nmi->in_nmi); in uv_check_nmi() 250 nmi = atomic_read(&uv_in_nmi); in uv_check_nmi() 268 if (cpu == atomic_read(&hub_nmi->cpu_owner)) { in uv_clear_nmi() 376 atomic_read(&uv_nmi_cpus_in_nmi), num_online_cpus()); in uv_nmi_wait() 440 while (atomic_read(&uv_nmi_cpus_in_nmi) > 0) in uv_nmi_sync_exit() 444 while (atomic_read(&uv_nmi_slave_continue)) in uv_nmi_sync_exit() 459 atomic_read(&uv_nmi_cpus_in_nmi), cpu); in uv_nmi_dump_state() 478 while (!atomic_read(&uv_nmi_slave_continue)) in uv_nmi_dump_state() 515 while (atomic_read(&uv_nmi_kexec_failed) == 0) in uv_nmi_kdump() [all …]
|
/linux-4.1.27/fs/ |
D | mbcache.c | 195 mb_assert(!(ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt))); in __mb_cache_entry_forget() 215 if (!(ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt))) { in __mb_cache_entry_release() 263 if (ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt)) in mb_cache_shrink_scan() 270 if (ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt) || in mb_cache_shrink_scan() 299 atomic_read(&cache->c_entry_count)); in mb_cache_shrink_count() 300 count += atomic_read(&cache->c_entry_count); in mb_cache_shrink_count() 410 atomic_read(&ce->e_refcnt)) in mb_cache_shrink() 420 atomic_read(&ce->e_refcnt) || in mb_cache_shrink() 430 atomic_read(&ce->e_refcnt))); in mb_cache_shrink() 473 atomic_read(&ce->e_refcnt))); in mb_cache_destroy() [all …]
|
/linux-4.1.27/fs/xfs/ |
D | xfs_trans_buf.c | 164 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_get_buf_map() 215 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_getsb() 285 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_read_buf_map() 374 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_brelse() 435 ASSERT(atomic_read(&bip->bli_refcount) == 0); in xfs_trans_brelse() 461 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_bhold() 481 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_bhold_release() 523 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_log_buf() 594 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_binval() 648 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_inode_buf() [all …]
|
D | xfs_dquot_item.c | 104 ASSERT(atomic_read(&dqp->q_pincount) > 0); in xfs_qm_dquot_logitem_unpin() 130 if (atomic_read(&dqp->q_pincount) == 0) in xfs_qm_dqunpin_wait() 137 wait_event(dqp->q_pinwait, (atomic_read(&dqp->q_pincount) == 0)); in xfs_qm_dqunpin_wait() 151 if (atomic_read(&dqp->q_pincount) > 0) in xfs_qm_dquot_logitem_push() 161 if (atomic_read(&dqp->q_pincount) > 0) { in xfs_qm_dquot_logitem_push()
|
/linux-4.1.27/fs/cifs/ |
D | smb2ops.c | 469 atomic_read(&sent[SMB2_NEGOTIATE_HE]), in smb2_print_stats() 470 atomic_read(&failed[SMB2_NEGOTIATE_HE])); in smb2_print_stats() 472 atomic_read(&sent[SMB2_SESSION_SETUP_HE]), in smb2_print_stats() 473 atomic_read(&failed[SMB2_SESSION_SETUP_HE])); in smb2_print_stats() 475 atomic_read(&sent[SMB2_LOGOFF_HE]), in smb2_print_stats() 476 atomic_read(&failed[SMB2_LOGOFF_HE])); in smb2_print_stats() 478 atomic_read(&sent[SMB2_TREE_CONNECT_HE]), in smb2_print_stats() 479 atomic_read(&failed[SMB2_TREE_CONNECT_HE])); in smb2_print_stats() 481 atomic_read(&sent[SMB2_TREE_DISCONNECT_HE]), in smb2_print_stats() 482 atomic_read(&failed[SMB2_TREE_DISCONNECT_HE])); in smb2_print_stats() [all …]
|
D | smb1ops.c | 654 atomic_read(&tcon->stats.cifs_stats.num_oplock_brks)); in cifs_print_stats() 656 atomic_read(&tcon->stats.cifs_stats.num_reads), in cifs_print_stats() 659 atomic_read(&tcon->stats.cifs_stats.num_writes), in cifs_print_stats() 662 atomic_read(&tcon->stats.cifs_stats.num_flushes)); in cifs_print_stats() 664 atomic_read(&tcon->stats.cifs_stats.num_locks), in cifs_print_stats() 665 atomic_read(&tcon->stats.cifs_stats.num_hardlinks), in cifs_print_stats() 666 atomic_read(&tcon->stats.cifs_stats.num_symlinks)); in cifs_print_stats() 668 atomic_read(&tcon->stats.cifs_stats.num_opens), in cifs_print_stats() 669 atomic_read(&tcon->stats.cifs_stats.num_closes), in cifs_print_stats() 670 atomic_read(&tcon->stats.cifs_stats.num_deletes)); in cifs_print_stats() [all …]
|
D | cifs_debug.c | 180 atomic_read(&server->in_send), in cifs_debug_data_proc_show() 181 atomic_read(&server->num_waiters)); in cifs_debug_data_proc_show() 318 atomic_read(&totBufAllocCount), in cifs_stats_proc_show() 319 atomic_read(&totSmBufAllocCount)); in cifs_stats_proc_show() 322 seq_printf(m, "Operations (MIDs): %d\n", atomic_read(&midCount)); in cifs_stats_proc_show() 348 atomic_read(&tcon->num_smbs_sent)); in cifs_stats_proc_show()
|
/linux-4.1.27/arch/x86/kernel/ |
D | tsc_sync.c | 149 while (atomic_read(&start_count) != cpus-1) in check_tsc_sync_source() 158 while (atomic_read(&stop_count) != cpus-1) in check_tsc_sync_source() 201 while (atomic_read(&start_count) != cpus) in check_tsc_sync_target() 214 while (atomic_read(&stop_count) != cpus) in check_tsc_sync_target()
|
/linux-4.1.27/include/asm-generic/ |
D | qrwlock.h | 48 return !(atomic_read(&lock->cnts) & _QW_WMASK); in queue_read_can_lock() 57 return !atomic_read(&lock->cnts); in queue_write_can_lock() 69 cnts = atomic_read(&lock->cnts); in queue_read_trylock() 88 cnts = atomic_read(&lock->cnts); in queue_write_trylock()
|
D | atomic.h | 128 #ifndef atomic_read 129 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 181 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/fs/gfs2/ |
D | log.c | 315 gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= in gfs2_log_release() 353 free_blocks = atomic_read(&sdp->sd_log_blks_free); in gfs2_log_reserve() 360 if (atomic_read(&sdp->sd_log_blks_free) <= wanted) in gfs2_log_reserve() 362 free_blocks = atomic_read(&sdp->sd_log_blks_free); in gfs2_log_reserve() 494 gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= in log_pull_tail() 505 if (atomic_read(&sdp->sd_log_in_flight)) { in log_flush_wait() 509 if (atomic_read(&sdp->sd_log_in_flight)) in log_flush_wait() 511 } while(atomic_read(&sdp->sd_log_in_flight)); in log_flush_wait() 662 enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); in log_write_header() 708 enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); in gfs2_log_flush() [all …]
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
D | i915_gem_debug.c | 46 !atomic_read(&obj->base.refcount.refcount)) { in i915_verify_lists() 68 !atomic_read(&obj->base.refcount.refcount)) { in i915_verify_lists() 86 !atomic_read(&obj->base.refcount.refcount)) { in i915_verify_lists() 102 !atomic_read(&obj->base.refcount.refcount)) { in i915_verify_lists()
|
/linux-4.1.27/drivers/md/ |
D | faulty.c | 97 atomic_read(&conf->counters[mode]) <= 0) in check_mode() 180 if (atomic_read(&conf->counters[WriteAll])) { in make_request() 234 if ((n=atomic_read(&conf->counters[WriteTransient])) != 0) in status() 238 if ((n=atomic_read(&conf->counters[ReadTransient])) != 0) in status() 242 if ((n=atomic_read(&conf->counters[WritePersistent])) != 0) in status() 246 if ((n=atomic_read(&conf->counters[ReadPersistent])) != 0) in status() 251 if ((n=atomic_read(&conf->counters[ReadFixable])) != 0) in status() 255 if ((n=atomic_read(&conf->counters[WriteAll])) != 0) in status()
|
D | dm-queue-length.c | 98 DMEMIT("%d ", atomic_read(&pi->qlen)); in ql_status() 185 (atomic_read(&pi->qlen) < atomic_read(&best->qlen))) in ql_select_path() 188 if (!atomic_read(&best->qlen)) in ql_select_path()
|
D | dm-service-time.c | 93 DMEMIT("%d %u ", atomic_read(&pi->in_flight_size), in st_status() 201 sz1 = atomic_read(&pi1->in_flight_size); in st_compare_load() 202 sz2 = atomic_read(&pi2->in_flight_size); in st_compare_load()
|
/linux-4.1.27/arch/m68k/include/asm/ |
D | atomic.h | 20 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 52 : "g" (i), "2" (atomic_read(v))); \ 131 prev = atomic_read(v); in atomic_cmpxchg() 144 prev = atomic_read(v); in atomic_xchg() 186 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/drivers/s390/char/ |
D | monreader.c | 212 if (!atomic_read(&monpriv->read_ready)) in mon_next_message() 312 atomic_read(&monpriv->iucv_connected) || in mon_open() 313 atomic_read(&monpriv->iucv_severed)); in mon_open() 314 if (atomic_read(&monpriv->iucv_severed)) { in mon_open() 381 atomic_read(&monpriv->read_ready) || in mon_read() 382 atomic_read(&monpriv->iucv_severed)); in mon_read() 385 if (unlikely(atomic_read(&monpriv->iucv_severed))) in mon_read() 436 if (unlikely(atomic_read(&monpriv->iucv_severed))) in mon_poll() 438 if (atomic_read(&monpriv->read_ready)) in mon_poll() 505 atomic_read(&monpriv->iucv_connected) || in monreader_thaw() [all …]
|
/linux-4.1.27/arch/sh/kernel/ |
D | ftrace.c | 108 atomic_read(&nmi_update_count)); in ftrace_arch_read_dyn_info() 114 int old = atomic_read(&nmi_running); in clear_mod_flag() 174 if (!atomic_read(&nmi_running)) in wait_for_nmi() 179 } while (atomic_read(&nmi_running)); in wait_for_nmi() 350 if (unlikely(atomic_read(¤t->tracing_graph_pause))) in prepare_ftrace_return()
|
D | smp.c | 362 if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) { in flush_tlb_mm() 394 if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) { in flush_tlb_range() 437 if ((atomic_read(&vma->vm_mm->mm_users) != 1) || in flush_tlb_page()
|
/linux-4.1.27/fs/ocfs2/dlm/ |
D | dlmdebug.c | 84 atomic_read(&lock->lock_refs.refcount), in __dlm_print_lock() 109 res->last_used, atomic_read(&res->refs.refcount), in __dlm_print_one_lock_resource() 117 res->inflight_locks, atomic_read(&res->asts_reserved)); in __dlm_print_one_lock_resource() 301 atomic_read(&mle->mle_refs.refcount)); in dump_mle() 517 atomic_read(&lock->lock_refs.refcount)); in dump_lock() 543 atomic_read(&res->asts_reserved), in dump_lockres() 544 atomic_read(&res->refs.refcount)); in dump_lockres() 757 atomic_read(&dlm->res_cur_count), in debug_state_print() 758 atomic_read(&dlm->res_tot_count)); in debug_state_print() 761 tot_mles += atomic_read(&dlm->mle_tot_count[i]); in debug_state_print() [all …]
|
/linux-4.1.27/kernel/debug/ |
D | debug_core.c | 401 if (atomic_read(&kgdb_setting_breakpoint)) in kgdb_io_ready() 418 if (atomic_read(&kgdb_active) != raw_smp_processor_id()) in kgdb_reenter_check() 555 if (atomic_read(&kgdb_cpu_doing_single_step) != -1 && in kgdb_cpu_enter() 603 (atomic_read(&masters_in_kgdb) + atomic_read(&slaves_in_kgdb)) != in kgdb_cpu_enter() 652 while (kgdb_do_roundup && atomic_read(&slaves_in_kgdb)) in kgdb_cpu_enter() 657 if (atomic_read(&kgdb_cpu_doing_single_step) != -1) { in kgdb_cpu_enter() 658 int sstep_cpu = atomic_read(&kgdb_cpu_doing_single_step); in kgdb_cpu_enter() 794 if (!kgdb_connected || atomic_read(&kgdb_active) != -1 || dbg_kdb_mode) in kgdb_console_write() 959 if (atomic_read(&kgdb_break_tasklet_var) || in kgdb_schedule_breakpoint() 960 atomic_read(&kgdb_active) != -1 || in kgdb_schedule_breakpoint() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/ |
D | socklnd.h | 419 LASSERT(atomic_read(&conn->ksnc_conn_refcount) > 0); in ksocknal_conn_addref() 429 LASSERT(atomic_read(&conn->ksnc_conn_refcount) > 0); in ksocknal_conn_decref() 441 LASSERT(atomic_read(&conn->ksnc_sock_refcount) > 0); in ksocknal_connsock_addref() 453 LASSERT(atomic_read(&conn->ksnc_sock_refcount) > 0); in ksocknal_connsock_decref() 465 LASSERT(atomic_read(&tx->tx_refcount) > 0); in ksocknal_tx_addref() 475 LASSERT(atomic_read(&tx->tx_refcount) > 0); in ksocknal_tx_decref() 483 LASSERT(atomic_read(&route->ksnr_refcount) > 0); in ksocknal_route_addref() 492 LASSERT(atomic_read(&route->ksnr_refcount) > 0); in ksocknal_route_decref() 500 LASSERT(atomic_read(&peer->ksnp_refcount) > 0); in ksocknal_peer_addref() 509 LASSERT(atomic_read(&peer->ksnp_refcount) > 0); in ksocknal_peer_decref()
|
/linux-4.1.27/arch/powerpc/platforms/powermac/ |
D | backlight.c | 102 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_key_worker() 130 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_key() 169 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_set_legacy_worker() 177 if (atomic_read(&kernel_backlight_disabled)) in pmac_backlight_set_legacy_brightness_pmu()
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
D | qxl_debugfs.c | 45 seq_printf(m, "%d\n", atomic_read(&qdev->irq_received)); in qxl_debugfs_irq_received() 46 seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_display)); in qxl_debugfs_irq_received() 47 seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_cursor)); in qxl_debugfs_irq_received() 48 seq_printf(m, "%d\n", atomic_read(&qdev->irq_received_io_cmd)); in qxl_debugfs_irq_received()
|
D | qxl_cmd.c | 288 irq_num = atomic_read(&qdev->irq_received_io_cmd); in wait_for_io_cmd_user() 292 atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); in wait_for_io_cmd_user() 295 atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); in wait_for_io_cmd_user() 299 irq_num = atomic_read(&qdev->irq_received_io_cmd); in wait_for_io_cmd_user() 305 atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); in wait_for_io_cmd_user() 308 atomic_read(&qdev->irq_received_io_cmd) > irq_num, 5*HZ); in wait_for_io_cmd_user()
|
/linux-4.1.27/drivers/mtd/maps/ |
D | vmu-flash.c | 149 if (atomic_read(&mdev->busy) == 1) { in maple_vmu_read_block() 151 atomic_read(&mdev->busy) == 0, HZ); in maple_vmu_read_block() 152 if (atomic_read(&mdev->busy) == 1) { in maple_vmu_read_block() 175 (atomic_read(&mdev->busy) == 0 || in maple_vmu_read_block() 176 atomic_read(&mdev->busy) == 2), HZ * 3); in maple_vmu_read_block() 182 if (error || atomic_read(&mdev->busy) == 2) { in maple_vmu_read_block() 183 if (atomic_read(&mdev->busy) == 2) in maple_vmu_read_block() 254 if (atomic_read(&mdev->busy) == 1) { in maple_vmu_write_block() 256 atomic_read(&mdev->busy) == 0, HZ); in maple_vmu_write_block() 257 if (atomic_read(&mdev->busy) == 1) { in maple_vmu_write_block() [all …]
|
/linux-4.1.27/drivers/connector/ |
D | cn_proc.c | 74 if (atomic_read(&proc_event_num_listeners) < 1) in proc_fork_connector() 105 if (atomic_read(&proc_event_num_listeners) < 1) in proc_exec_connector() 131 if (atomic_read(&proc_event_num_listeners) < 1) in proc_id_connector() 169 if (atomic_read(&proc_event_num_listeners) < 1) in proc_sid_connector() 194 if (atomic_read(&proc_event_num_listeners) < 1) in proc_ptrace_connector() 227 if (atomic_read(&proc_event_num_listeners) < 1) in proc_comm_connector() 253 if (atomic_read(&proc_event_num_listeners) < 1) in proc_coredump_connector() 278 if (atomic_read(&proc_event_num_listeners) < 1) in proc_exit_connector() 313 if (atomic_read(&proc_event_num_listeners) < 1) in cn_proc_ack()
|
D | cn_queue.c | 153 while (atomic_read(&dev->refcnt)) { in cn_queue_free_dev() 155 dev->name, atomic_read(&dev->refcnt)); in cn_queue_free_dev()
|
/linux-4.1.27/net/atm/ |
D | atm_misc.c | 17 if (atomic_read(&sk_atm(vcc)->sk_rmem_alloc) <= sk_atm(vcc)->sk_rcvbuf) in atm_charge() 32 if (atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf) { in atm_alloc_charge() 89 #define __HANDLE_ITEM(i) to->i = atomic_read(&from->i) in sonet_copy_stats()
|
D | proc.c | 48 atomic_read(&stats->tx), atomic_read(&stats->tx_err), in add_stats() 49 atomic_read(&stats->rx), atomic_read(&stats->rx_err), in add_stats() 50 atomic_read(&stats->rx_drop)); in add_stats() 64 seq_printf(seq, "\t[%d]", atomic_read(&dev->refcnt)); in atm_dev_info() 214 atomic_read(&sk->sk_refcnt)); in vcc_info()
|
/linux-4.1.27/lib/ |
D | fault-inject.c | 48 atomic_read(&attr->space), in fail_dump() 49 atomic_read(&attr->times)); in fail_dump() 117 if (atomic_read(&attr->times) == 0) in should_fail() 120 if (atomic_read(&attr->space) > size) { in should_fail() 139 if (atomic_read(&attr->times) != -1) in should_fail()
|
D | is_single_threaded.c | 25 if (atomic_read(&task->signal->live) != 1) in current_is_single_threaded() 28 if (atomic_read(&mm->mm_users) == 1) in current_is_single_threaded()
|
/linux-4.1.27/fs/coda/ |
D | cache.c | 35 cii->c_cached_epoch = atomic_read(&permission_epoch); in coda_cache_enter() 49 cii->c_cached_epoch = atomic_read(&permission_epoch) - 1; in coda_cache_clear_inode() 69 cii->c_cached_epoch == atomic_read(&permission_epoch); in coda_cache_check()
|
/linux-4.1.27/fs/cachefiles/ |
D | proc.c | 38 x = atomic_read(&cachefiles_lookup_histogram[index]); in cachefiles_histogram_show() 39 y = atomic_read(&cachefiles_mkdir_histogram[index]); in cachefiles_histogram_show() 40 z = atomic_read(&cachefiles_create_histogram[index]); in cachefiles_histogram_show()
|
D | interface.c | 185 _enter("{OBJ%x,%d}", _object->debug_id, atomic_read(&object->usage)); in cachefiles_grab_object() 188 ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); in cachefiles_grab_object() 262 object->fscache.debug_id, atomic_read(&object->usage)); in cachefiles_drop_object() 268 ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); in cachefiles_drop_object() 323 object->fscache.debug_id, atomic_read(&object->usage)); in cachefiles_put_object() 326 ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); in cachefiles_put_object()
|
/linux-4.1.27/drivers/staging/lustre/lustre/ldlm/ |
D | ldlm_pool.c | 226 granted = atomic_read(&pl->pl_granted); in ldlm_pool_recalc_grant_plan() 254 granted = atomic_read(&pl->pl_granted); in ldlm_pool_recalc_slv() 289 int granted = atomic_read(&pl->pl_granted); in ldlm_pool_recalc_stats() 290 int grant_rate = atomic_read(&pl->pl_grant_rate); in ldlm_pool_recalc_stats() 291 int cancel_rate = atomic_read(&pl->pl_cancel_rate); in ldlm_pool_recalc_stats() 385 return atomic_read(&pl->pl_granted); in ldlm_srv_pool_shrink() 391 if (atomic_read(&pl->pl_granted) == 0) in ldlm_srv_pool_shrink() 671 granted = atomic_read(&pl->pl_granted); in lprocfs_pool_state_seq_show() 672 grant_rate = atomic_read(&pl->pl_grant_rate); in lprocfs_pool_state_seq_show() 673 cancel_rate = atomic_read(&pl->pl_cancel_rate); in lprocfs_pool_state_seq_show() [all …]
|
D | ldlm_resource.c | 770 atomic_read(&res->lr_refcount) - 1); in ldlm_resource_complain() 807 if (atomic_read(&ns->ns_bref) > 0) { in __ldlm_namespace_free() 813 ldlm_ns_name(ns), atomic_read(&ns->ns_bref)); in __ldlm_namespace_free() 819 atomic_read(&ns->ns_bref) == 0, &lwi); in __ldlm_namespace_free() 826 atomic_read(&ns->ns_bref), rc); in __ldlm_namespace_free() 830 if (atomic_read(&ns->ns_bref)) { in __ldlm_namespace_free() 833 atomic_read(&ns->ns_bref), rc); in __ldlm_namespace_free() 1174 atomic_read(&res->lr_refcount)); in ldlm_resource_getref() 1213 res, atomic_read(&res->lr_refcount) - 1); in ldlm_resource_putref() 1235 res, atomic_read(&res->lr_refcount) - 1); in ldlm_resource_putref_locked() [all …]
|
D | ldlm_lock.c | 189 LASSERT(atomic_read(&lock->l_refc) > 0); in ldlm_lock_put() 1857 rc = atomic_read(&arg->restart) ? -ERESTART : 0; in ldlm_run_ast_work() 2042 atomic_read(&hs->hs_count)); in ldlm_cancel_locks_for_export_cb() 2232 lock->l_handle.h_cookie, atomic_read(&lock->l_refc), in _ldlm_lock_debug() 2237 exp ? atomic_read(&exp->exp_refcount) : -99, in _ldlm_lock_debug() 2248 lock->l_handle.h_cookie, atomic_read(&lock->l_refc), in _ldlm_lock_debug() 2253 atomic_read(&resource->lr_refcount), in _ldlm_lock_debug() 2259 exp ? atomic_read(&exp->exp_refcount) : -99, in _ldlm_lock_debug() 2268 lock->l_handle.h_cookie, atomic_read(&lock->l_refc), in _ldlm_lock_debug() 2273 atomic_read(&resource->lr_refcount), in _ldlm_lock_debug() [all …]
|
/linux-4.1.27/drivers/md/bcache/ |
D | writeback.h | 12 ret += atomic_read(d->stripe_sectors_dirty + i); in bcache_dev_sectors_dirty() 31 if (atomic_read(dc->disk.stripe_sectors_dirty + stripe)) in bcache_dev_stripe_dirty() 72 if (!atomic_read(&dc->has_dirty) && in bch_writeback_add()
|
D | closure.c | 103 if (atomic_read(&cl->remaining) & CLOSURE_WAITING) in closure_wait() 126 if ((atomic_read(&cl->remaining) & in closure_sync() 178 int r = atomic_read(&cl->remaining); in debug_seq_show()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/core/ |
D | object.c | 147 nv_trace(obj, "inc() == %d\n", atomic_read(&obj->refcount)); in nvkm_object_ref() 152 nv_trace(*ref, "dec() == %d\n", atomic_read(&(*ref)->refcount)); in nvkm_object_ref() 166 nv_trace(object, "use(+1) == %d\n", atomic_read(&object->usecount)); in nvkm_object_inc() 296 nv_trace(object, "use(-1) == %d\n", atomic_read(&object->usecount)); in nvkm_object_dec() 325 atomic_read(&object->refcount), in nvkm_object_debug() 326 atomic_read(&object->usecount)); in nvkm_object_debug()
|
/linux-4.1.27/drivers/net/wireless/mwifiex/ |
D | main.c | 182 (atomic_read(&adapter->rx_pending) < LOW_RX_PENDING)) { in mwifiex_process_rx() 250 if (atomic_read(&adapter->rx_pending) >= HIGH_RX_PENDING && in mwifiex_main_process() 891 atomic_read(&cardp->tx_cmd_urb_pending)); in mwifiex_dump_drv_info() 893 atomic_read(&cardp->tx_data_urb_pending)); in mwifiex_dump_drv_info() 895 atomic_read(&cardp->rx_cmd_urb_pending)); in mwifiex_dump_drv_info() 897 atomic_read(&cardp->rx_data_urb_pending)); in mwifiex_dump_drv_info() 901 atomic_read(&adapter->tx_pending)); in mwifiex_dump_drv_info() 903 atomic_read(&adapter->rx_pending)); in mwifiex_dump_drv_info() 920 atomic_read(&priv->wmm_tx_pending[0])); in mwifiex_dump_drv_info() 922 atomic_read(&priv->wmm_tx_pending[1])); in mwifiex_dump_drv_info() [all …]
|
D | usb.c | 239 if (atomic_read(&adapter->rx_pending) <= HIGH_RX_PENDING) in mwifiex_usb_rx_complete() 313 if (atomic_read(&card->rx_cmd_urb_pending) && card->rx_cmd.urb) in mwifiex_usb_free() 319 if (atomic_read(&card->rx_data_urb_pending)) in mwifiex_usb_free() 484 if (atomic_read(&card->rx_cmd_urb_pending) && card->rx_cmd.urb) in mwifiex_usb_suspend() 487 if (atomic_read(&card->rx_data_urb_pending)) in mwifiex_usb_suspend() 531 if (!atomic_read(&card->rx_data_urb_pending)) in mwifiex_usb_resume() 536 if (!atomic_read(&card->rx_cmd_urb_pending)) { in mwifiex_usb_resume() 736 atomic_read(&card->tx_data_urb_pending) >= MWIFIEX_TX_DATA_URB) { in mwifiex_usb_host_to_card() 781 atomic_read(&card->tx_data_urb_pending) == in mwifiex_usb_host_to_card() 989 (!atomic_read(&card->rx_cmd_urb_pending))) in mwifiex_submit_rx_urb()
|
D | debugfs.c | 295 atomic_read(&phist_data->num_samples)); in mwifiex_histogram_read() 311 value = atomic_read(&phist_data->rx_rate[i]); in mwifiex_histogram_read() 319 value = atomic_read(&phist_data->rx_rate[i]); in mwifiex_histogram_read() 327 value = atomic_read(&phist_data->snr[i]); in mwifiex_histogram_read() 332 value = atomic_read(&phist_data->noise_flr[i]); in mwifiex_histogram_read() 338 value = atomic_read(&phist_data->sig_str[i]); in mwifiex_histogram_read()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | crash.c | 124 while ((atomic_read(&cpus_in_crash) < ncpus) && (--msecs > 0)) in crash_kexec_prepare_cpus() 129 if (atomic_read(&cpus_in_crash) >= ncpus) { in crash_kexec_prepare_cpus() 135 ncpus - atomic_read(&cpus_in_crash)); in crash_kexec_prepare_cpus() 166 while (atomic_read(&cpus_in_crash) < ncpus) in crash_kexec_prepare_cpus()
|
D | rtas.c | 737 while (rc == H_MULTI_THREADS_ACTIVE && !atomic_read(&data->done) && in __rtas_suspend_last_cpu() 738 !atomic_read(&data->error)) in __rtas_suspend_last_cpu() 741 if (rc || atomic_read(&data->error)) { in __rtas_suspend_last_cpu() 746 if (atomic_read(&data->error)) in __rtas_suspend_last_cpu() 747 rc = atomic_read(&data->error); in __rtas_suspend_last_cpu() 783 while (rc == H_SUCCESS && !atomic_read(&data->done) && !atomic_read(&data->error)) in __rtas_suspend_cpu() 974 if (atomic_read(&data.error) != 0) in rtas_ibm_suspend_me() 987 return atomic_read(&data.error); in rtas_ibm_suspend_me()
|
/linux-4.1.27/drivers/block/drbd/ |
D | drbd_proc.c | 317 atomic_read(&device->local_cnt), in drbd_seq_show() 318 atomic_read(&device->ap_pending_cnt) + in drbd_seq_show() 319 atomic_read(&device->rs_pending_cnt), in drbd_seq_show() 320 atomic_read(&device->unacked_cnt), in drbd_seq_show() 321 atomic_read(&device->ap_bio_cnt), in drbd_seq_show() 342 seq_printf(seq, "\tblocked on activity log: %d\n", atomic_read(&device->ap_actlog_cnt)); in drbd_seq_show()
|
/linux-4.1.27/sound/usb/usx2y/ |
D | usbusx2yaudio.c | 151 if (atomic_read(&subs->state) >= state_PRERUNNING) in usX2Y_urb_play_prepare() 219 state = atomic_read(&playbacksubs->state); in usX2Y_usbframe_complete() 246 state = atomic_read(&capsubs->state); in usX2Y_usbframe_complete() 268 snd_printdd("%i %p state=%i\n", s, subs, atomic_read(&subs->state)); in usX2Y_clients_stop() 275 if (atomic_read(&subs->state) >= state_PRERUNNING) in usX2Y_clients_stop() 302 if (unlikely(atomic_read(&subs->state) < state_PREPARED)) { in i_usX2Y_urb_complete() 320 atomic_read(&capsubs->state) >= state_PREPARED && in i_usX2Y_urb_complete() 322 atomic_read(&playbacksubs->state) < state_PREPARED)) { in i_usX2Y_urb_complete() 476 if (subs != NULL && atomic_read(&subs->state) >= state_PREPARED) in usX2Y_urbs_start() 509 if (atomic_read(&subs->state) != state_PREPARED) in usX2Y_urbs_start() [all …]
|
D | usx2yhwdeppcm.c | 137 if (atomic_read(&subs->state) != state_RUNNING) in usX2Y_hwdep_urb_play_prepare() 180 state = atomic_read(&playbacksubs->state); in usX2Y_usbpcm_usbframe_complete() 207 state = atomic_read(&capsubs->state); in usX2Y_usbpcm_usbframe_complete() 236 if (unlikely(atomic_read(&subs->state) < state_PREPARED)) { in i_usX2Y_usbpcm_urb_complete() 252 if (capsubs->completed_urb && atomic_read(&capsubs->state) >= state_PREPARED && in i_usX2Y_usbpcm_urb_complete() 254 (playbacksubs->completed_urb || atomic_read(&playbacksubs->state) < state_PREPARED)) { in i_usX2Y_usbpcm_urb_complete() 381 if (atomic_read(&playback_subs->state) < state_PREPARED) { in snd_usX2Y_usbpcm_hw_free() 425 if (subs != NULL && atomic_read(&subs->state) >= state_PREPARED) in usX2Y_usbpcm_urbs_start() 465 if (atomic_read(&subs->state) != state_PREPARED) in usX2Y_usbpcm_urbs_start() 500 if (atomic_read(&capsubs->state) < state_PREPARED) { in snd_usX2Y_usbpcm_prepare() [all …]
|
/linux-4.1.27/drivers/gpu/drm/via/ |
D | via_irq.c | 104 return atomic_read(&dev_priv->vbl_received); in via_get_vblank_counter() 120 if (!(atomic_read(&dev_priv->vbl_received) & 0x0F)) { in via_driver_irq_handler() 130 if (!(atomic_read(&dev_priv->vbl_received) & 0xFF)) { in via_driver_irq_handler() 245 cur_irq_sequence = atomic_read(&cur_irq->irq_received); in via_driver_irq_wait() 249 atomic_read(&cur_irq->irq_received)) - in via_driver_irq_wait() 369 atomic_read(&cur_irq->irq_received); in via_wait_irq()
|
/linux-4.1.27/drivers/cpuidle/ |
D | coupled.c | 165 while (atomic_read(a) < n) in cpuidle_coupled_parallel_barrier() 173 while (atomic_read(a) > n) in cpuidle_coupled_parallel_barrier() 233 int r = atomic_read(&coupled->ready_waiting_counts) >> WAITING_BITS; in cpuidle_coupled_no_cpus_ready() 245 int r = atomic_read(&coupled->ready_waiting_counts) >> WAITING_BITS; in cpuidle_coupled_cpus_ready() 257 int w = atomic_read(&coupled->ready_waiting_counts) & WAITING_MASK; in cpuidle_coupled_cpus_waiting() 269 int w = atomic_read(&coupled->ready_waiting_counts) & WAITING_MASK; in cpuidle_coupled_no_cpus_waiting()
|
/linux-4.1.27/drivers/net/wireless/cw1200/ |
D | bh.c | 141 (CW1200_BH_SUSPENDED == atomic_read(&priv->bh_suspend)), in cw1200_bh_suspend() 156 (CW1200_BH_RESUMED == atomic_read(&priv->bh_suspend)), in cw1200_bh_resume() 439 !atomic_read(&priv->recent_scan)) { in cw1200_bh() 453 (atomic_read(&priv->bh_rx) == 0) && in cw1200_bh() 454 (atomic_read(&priv->bh_tx) == 0)) in cw1200_bh() 464 0 : atomic_read(&priv->bh_suspend); in cw1200_bh() 514 !atomic_read(&priv->recent_scan)) { in cw1200_bh() 533 CW1200_BH_RESUME == atomic_read(&priv->bh_suspend)); in cw1200_bh()
|
D | debug.c | 235 atomic_read(&priv->bh_term) ? "terminated" : "alive"); in cw1200_status_show() 237 atomic_read(&priv->bh_rx)); in cw1200_status_show() 239 atomic_read(&priv->bh_tx)); in cw1200_status_show() 263 atomic_read(&priv->tx_lock) ? "locked" : "unlocked"); in cw1200_status_show() 264 if (atomic_read(&priv->tx_lock)) in cw1200_status_show() 266 atomic_read(&priv->tx_lock)); in cw1200_status_show() 287 atomic_read(&priv->scan.in_progress) ? "active" : "idle"); in cw1200_status_show()
|
/linux-4.1.27/net/nfc/nci/ |
D | core.c | 622 if ((atomic_read(&ndev->state) == NCI_DISCOVERY) || in nci_start_poll() 623 (atomic_read(&ndev->state) == NCI_W4_ALL_DISCOVERIES)) { in nci_start_poll() 633 if ((atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) || in nci_start_poll() 634 (atomic_read(&ndev->state) == NCI_POLL_ACTIVE)) { in nci_start_poll() 673 if ((atomic_read(&ndev->state) != NCI_DISCOVERY) && in nci_stop_poll() 674 (atomic_read(&ndev->state) != NCI_W4_ALL_DISCOVERIES)) { in nci_stop_poll() 694 if ((atomic_read(&ndev->state) != NCI_W4_HOST_SELECT) && in nci_activate_target() 695 (atomic_read(&ndev->state) != NCI_POLL_ACTIVE)) { in nci_activate_target() 723 if (atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) { in nci_activate_target() 763 if (atomic_read(&ndev->state) == NCI_POLL_ACTIVE) { in nci_deactivate_target() [all …]
|
/linux-4.1.27/drivers/ras/ |
D | debugfs.c | 9 return atomic_read(&trace_count); in ras_userspace_consumers() 15 return atomic_read(&trace_count); in trace_show()
|
/linux-4.1.27/drivers/staging/lustre/lustre/lov/ |
D | lov_internal.h | 129 LASSERT(atomic_read(&set->set_refcount) > 0); in lov_get_reqset() 149 atomic_read(&llh->llh_refcount) - 1); in lov_llh_put() 150 LASSERT(atomic_read(&llh->llh_refcount) > 0 && in lov_llh_put() 151 atomic_read(&llh->llh_refcount) < 0x5a5a); in lov_llh_put() 156 if (atomic_read(&llh->llh_refcount)) in lov_llh_put() 302 LASSERT(atomic_read(&lsm->lsm_refc) > 0); in lsm_addref()
|
D | lov_request.c | 91 int completes = atomic_read(&set->set_completes); in lov_set_finished() 214 if (!atomic_read(&set->set_success)) in common_attr_done() 263 if (atomic_read(&set->set_completes)) in lov_fini_getattr_set() 354 if (atomic_read(&set->set_completes)) { in lov_fini_destroy_set() 435 if (atomic_read(&set->set_completes)) { in lov_fini_setattr_set() 597 if (atomic_read(&set->set_completes)) { in lov_fini_statfs_set() 599 atomic_read(&set->set_success)); in lov_fini_statfs_set() 679 success = atomic_read(&set->set_success); in cb_statfs_update() 706 atomic_read(&set->set_success)); in cb_statfs_update()
|
D | lov_object.c | 410 lsm->lsm_magic, atomic_read(&lsm->lsm_refc), in lov_print_raid0() 434 lsm->lsm_magic, atomic_read(&lsm->lsm_refc), in lov_print_released() 641 while (atomic_read(&lov->lo_active_ios) > 0) { in lov_layout_wait() 644 atomic_read(&lov->lo_active_ios)); in lov_layout_wait() 647 atomic_read(&lov->lo_active_ios) == 0, &lwi); in lov_layout_wait() 691 LASSERT(atomic_read(&lov->lo_active_ios) == 0); in lov_layout_change() 760 atomic_read(&lov->lo_active_ios) > 0) { in lov_conf_set() 783 if (atomic_read(&lov->lo_active_ios) > 0) { in lov_conf_set() 919 lsm, atomic_read(&lsm->lsm_refc), in lov_lsm_addref() 932 lsm, atomic_read(&lsm->lsm_refc), current); in lov_lsm_decref()
|
/linux-4.1.27/mm/ |
D | mmu_notifier.c | 235 BUG_ON(atomic_read(&mm->mm_users) <= 0); in do_mmu_notifier_register() 281 BUG_ON(atomic_read(&mm->mm_users) <= 0); in do_mmu_notifier_register() 334 BUG_ON(atomic_read(&mm->mm_count) <= 0); in mmu_notifier_unregister() 367 BUG_ON(atomic_read(&mm->mm_count) <= 0); in mmu_notifier_unregister() 387 BUG_ON(atomic_read(&mm->mm_count) <= 0); in mmu_notifier_unregister_no_release()
|
D | debug.c | 87 page, atomic_read(&page->_count), page_mapcount(page), in dump_page_badflags() 206 mm->pgd, atomic_read(&mm->mm_users), in dump_mm() 207 atomic_read(&mm->mm_count), in dump_mm()
|
D | internal.h | 46 VM_BUG_ON_PAGE(atomic_read(&page->_count), page); in set_page_refcounted() 64 VM_BUG_ON_PAGE(atomic_read(&page->first_page->_count) <= 0, page); in __get_page_tail_foll() 89 VM_BUG_ON_PAGE(atomic_read(&page->_count) <= 0, page); in get_page_foll()
|
/linux-4.1.27/drivers/edac/ |
D | edac_stub.c | 60 return atomic_read(&edac_handlers); in edac_handler_set() 88 if (atomic_read(&edac_subsys_valid)) in edac_get_sysfs_subsys()
|
/linux-4.1.27/sound/drivers/pcsp/ |
D | pcsp_lib.c | 30 if (atomic_read(&pcsp_chip.timer_active)) { in pcsp_call_pcm_elapsed() 134 if (!atomic_read(&chip->timer_active) || !chip->playback_substream) in pcsp_do_timer() 157 if (atomic_read(&chip->timer_active)) { in pcsp_start_playing() 178 if (!atomic_read(&chip->timer_active)) in pcsp_stop_playing() 317 if (atomic_read(&chip->timer_active)) { in snd_pcsp_playback_open()
|
/linux-4.1.27/drivers/media/usb/usbtv/ |
D | usbtv-audio.c | 66 if (atomic_read(&chip->snd_stream)) { in snd_usbtv_pcm_close() 131 if (!atomic_read(&chip->snd_stream)) in usbtv_audio_urb_received() 267 if (atomic_read(&usbtv->snd_stream) && usbtv->snd_bulk_urb) in usbtv_audio_suspend() 273 if (atomic_read(&usbtv->snd_stream) && usbtv->snd_bulk_urb) in usbtv_audio_resume() 281 if (atomic_read(&chip->snd_stream)) in snd_usbtv_trigger()
|
/linux-4.1.27/include/linux/ |
D | iocontext.h | 131 WARN_ON_ONCE(atomic_read(&ioc->active_ref) <= 0); in get_io_context_active() 140 WARN_ON_ONCE(atomic_read(&ioc->nr_tasks) <= 0); in ioc_task_link()
|
D | atomic.h | 100 c = atomic_read(v); in atomic_dec_if_positive() 121 old = atomic_read(v); in atomic_or()
|
/linux-4.1.27/drivers/net/wireless/ath/carl9170/ |
D | debug.c | 224 ar->fw.mem_blocks, atomic_read(&ar->mem_allocs)); in carl9170_debugfs_mem_usage_read() 227 atomic_read(&ar->mem_free_blocks), in carl9170_debugfs_mem_usage_read() 228 (atomic_read(&ar->mem_free_blocks) * ar->fw.mem_block_size) / 1024, in carl9170_debugfs_mem_usage_read() 675 atomic_read(&ar->pending_restarts)); in carl9170_debugfs_bug_read() 771 atomic_read(&ar->tx_anch_urbs)); 773 atomic_read(&ar->rx_anch_urbs)); 775 atomic_read(&ar->rx_work_urbs)); 777 atomic_read(&ar->rx_pool_urbs)); 780 atomic_read(&ar->tx_total_queued)); 782 atomic_read(&ar->tx_ampdu_scheduler)); [all …]
|
/linux-4.1.27/sound/firewire/fireworks/ |
D | fireworks_stream.c | 218 if ((atomic_read(&efw->playback_substreams) == 0) && in snd_efw_stream_start_duplex() 219 (atomic_read(&efw->capture_substreams) == 0)) in snd_efw_stream_start_duplex() 278 if (atomic_read(slave_substreams) > 0 && !amdtp_stream_running(slave)) { in snd_efw_stream_start_duplex() 310 if (atomic_read(slave_substreams) == 0) { in snd_efw_stream_stop_duplex() 313 if (atomic_read(master_substreams) == 0) in snd_efw_stream_stop_duplex()
|
/linux-4.1.27/drivers/media/pci/ivtv/ |
D | ivtv-firmware.c | 346 if (!res && !atomic_read(&itv->capturing) && in ivtv_firmware_check() 347 (!atomic_read(&itv->decoding) || in ivtv_firmware_check() 348 (atomic_read(&itv->decoding) < 2 && test_bit(IVTV_F_I_DEC_YUV, in ivtv_firmware_check() 377 if (res && !atomic_read(&itv->capturing) && in ivtv_firmware_check() 378 !atomic_read(&itv->decoding)) { in ivtv_firmware_check()
|
/linux-4.1.27/fs/ext3/ |
D | fsync.c | 84 commit_tid = atomic_read(&ei->i_datasync_tid); in ext3_sync_file() 86 commit_tid = atomic_read(&ei->i_sync_tid); in ext3_sync_file()
|
/linux-4.1.27/arch/sh/include/asm/ |
D | atomic.h | 17 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 53 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/block/ |
D | blk-mq-tag.c | 53 int old = atomic_read(index); in bt_index_atomic_inc() 79 wake_index = atomic_read(&bt->wake_index); in blk_mq_tag_wakeup_all() 132 users = atomic_read(&hctx->tags->active_queues); in hctx_may_queue() 140 return atomic_read(&hctx->nr_active) < depth; in hctx_may_queue() 248 wait_index = atomic_read(&hctx->wait_index); in bt_wait_ptr() 358 wake_index = atomic_read(&bt->wake_index); in bt_wake_ptr() 363 int o = atomic_read(&bt->wake_index); in bt_wake_ptr() 663 page += sprintf(page, "active_queues=%u\n", atomic_read(&tags->active_queues)); in blk_mq_tag_sysfs_show()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
D | iwch_cm.h | 58 ep, atomic_read(&((ep)->kref.refcount))); \ 59 WARN_ON(atomic_read(&((ep)->kref.refcount)) < 1); \ 65 ep, atomic_read(&((ep)->kref.refcount))); \
|
/linux-4.1.27/fs/reiserfs/ |
D | journal.c | 741 if (atomic_read(&nr_reiserfs_jh) <= 0) in reiserfs_free_jh() 905 atomic_read(&other_jl->j_older_commits_done)) in flush_older_commits() 925 if (atomic_read(&other_jl->j_commit_left) != 0) { in flush_older_commits() 957 if (atomic_read(&j->j_async_throttle)) { in reiserfs_async_progress_wait() 989 if (atomic_read(&jl->j_older_commits_done)) { in flush_commit_list() 1021 if (atomic_read(&jl->j_commit_left) <= 0) { in flush_commit_list() 1102 BUG_ON(atomic_read(&jl->j_commit_left) != 1); in flush_commit_list() 1364 if (atomic_read(&journal->j_wcount) != 0) { in flush_journal_list() 1366 atomic_read(&journal->j_wcount)); in flush_journal_list() 1384 if (atomic_read(&jl->j_nonzerolen) <= 0 && in flush_journal_list() [all …]
|
/linux-4.1.27/include/trace/events/ |
D | sock.h | 23 __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc); 51 __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);
|
D | rpm.h | 37 __entry->usage_count = atomic_read( 43 __entry->child_count = atomic_read(
|
/linux-4.1.27/drivers/net/wireless/rsi/ |
D | rsi_common.h | 41 (atomic_read(&event->event_condition) == 0)); in rsi_wait_event() 44 (atomic_read(&event->event_condition) == 0), in rsi_wait_event()
|
/linux-4.1.27/drivers/net/irda/ |
D | vlsi_ir.h | 684 t = atomic_read(&r->tail) & r->mask; in ring_last() 685 return (((t+1) & r->mask) == (atomic_read(&r->head) & r->mask)) ? NULL : &r->rd[t]; in ring_last() 698 h = atomic_read(&r->head) & r->mask; in ring_first() 699 return (h == (atomic_read(&r->tail) & r->mask)) ? NULL : &r->rd[h]; in ring_first()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
D | l2t.c | 222 if (!atomic_read(&d->nfree)) in alloc_l2e() 227 if (atomic_read(&e->refcnt) == 0) in alloc_l2e() 230 for (e = &d->l2tab[1]; atomic_read(&e->refcnt); ++e) ; in alloc_l2e() 266 if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */ in t3_l2e_free() 337 if (atomic_read(&e->refcnt) == 1) in t3_l2t_get() 420 if (atomic_read(&e->refcnt)) { in t3_l2t_update()
|
/linux-4.1.27/drivers/gpu/host1x/ |
D | syncpt.h | 98 min = atomic_read(&sp->min_val); in host1x_syncpt_idle() 99 max = atomic_read(&sp->max_val); in host1x_syncpt_idle()
|
D | syncpt.c | 283 current_val = (u32)atomic_read(&sp->min_val); in host1x_syncpt_is_expired() 284 future_val = (u32)atomic_read(&sp->max_val); in host1x_syncpt_is_expired() 414 return (u32)atomic_read(&sp->max_val); in host1x_syncpt_read_max() 424 return (u32)atomic_read(&sp->min_val); in host1x_syncpt_read_min()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | nmi.c | 56 if (atomic_read(&nmi_active)) { in touch_nmi_watchdog() 166 if (!atomic_read(&nmi_active)) in check_nmi_watchdog() 191 if (!atomic_read(&nmi_active)) { in check_nmi_watchdog()
|
/linux-4.1.27/net/ipv4/ |
D | inet_timewait_sock.c | 90 BUG_ON(refcnt >= atomic_read(&tw->tw_refcnt)); in inet_twsk_kill() 192 if (atomic_read(&dr->tw_count) >= dr->sysctl_max_tw_buckets) in inet_twsk_alloc() 304 atomic_read(&twsk_net(tw)->count)) in inet_twsk_purge() 311 atomic_read(&twsk_net(tw)->count))) { in inet_twsk_purge()
|
/linux-4.1.27/drivers/base/ |
D | dd.c | 281 int local_trigger_count = atomic_read(&deferred_trigger_count); in really_probe() 340 if (local_trigger_count != atomic_read(&deferred_trigger_count)) in really_probe() 374 atomic_read(&probe_count)); in driver_probe_done() 375 if (atomic_read(&probe_count)) in driver_probe_done() 387 wait_event(probe_waitqueue, atomic_read(&probe_count) == 0); in wait_for_device_probe()
|
/linux-4.1.27/fs/pstore/ |
D | ram_core.c | 42 return atomic_read(&prz->buffer->size); in buffer_size() 47 return atomic_read(&prz->buffer->start); in buffer_start() 57 old = atomic_read(&prz->buffer->start); in buffer_start_add_atomic() 72 if (atomic_read(&prz->buffer->size) == prz->buffer_size) in buffer_size_add_atomic() 76 old = atomic_read(&prz->buffer->size); in buffer_size_add_atomic() 94 old = atomic_read(&prz->buffer->start); in buffer_start_add_locked() 114 old = atomic_read(&prz->buffer->size); in buffer_size_add_locked()
|
/linux-4.1.27/drivers/media/pci/cx18/ |
D | cx18-streams.c | 672 if ((atomic_read(&s->q_free.depth) + atomic_read(&s->q_busy.depth)) >= in cx18_stream_rotate_idx_mdls() 677 if (atomic_read(&s->q_full.depth) < 2) in cx18_stream_rotate_idx_mdls() 719 if (atomic_read(&s->q_free.depth) == 0 || in _cx18_stream_load_fw_queue() 720 atomic_read(&s->q_busy.depth) >= CX18_MAX_FW_MDLS_PER_STREAM) in _cx18_stream_load_fw_queue() 729 } while (atomic_read(&s->q_busy.depth) < CX18_MAX_FW_MDLS_PER_STREAM in _cx18_stream_load_fw_queue() 866 if (atomic_read(&cx->ana_capturing) == 0) in cx18_start_v4l2_encode_stream() 918 if (atomic_read(&cx->tot_capturing) == 0) { in cx18_start_v4l2_encode_stream() 947 if (atomic_read(&cx->tot_capturing) == 0) { in cx18_start_v4l2_encode_stream() 988 if (atomic_read(&cx->tot_capturing) == 0) in cx18_stop_v4l2_encode_stream() 1015 if (atomic_read(&cx->tot_capturing) > 0) in cx18_stop_v4l2_encode_stream()
|
D | cx18-fileops.c | 247 if (!atomic_read(&s->q_full.depth)) in cx18_get_mdl() 427 if (atomic_read(&cx->ana_capturing) == 0 && s->id == -1) { in cx18_read() 654 if (atomic_read(&s->q_full.depth)) in cx18_v4l2_enc_poll() 775 if (atomic_read(&cx->ana_capturing) > 0) { in cx18_v4l2_close() 821 if (atomic_read(&cx->ana_capturing) > 0) { in cx18_serialized_open() 867 if (atomic_read(&cx->ana_capturing)) { in cx18_mute() 880 if (atomic_read(&cx->ana_capturing)) { in cx18_unmute()
|
/linux-4.1.27/drivers/misc/sgi-xp/ |
D | xpc_main.c | 350 atomic_read(&part->nchannels_active) > 0 || in xpc_channel_mgr() 370 (atomic_read(&part->channel_mgr_requests) > 0 || in xpc_channel_mgr() 373 atomic_read(&part->nchannels_active) == 0 && in xpc_channel_mgr() 501 DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); in xpc_teardown_ch_structures() 502 DBUG_ON(atomic_read(&part->nchannels_active) != 0); in xpc_teardown_ch_structures() 512 wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); in xpc_teardown_ch_structures() 619 int idle = atomic_read(&ch->kthreads_idle); in xpc_activate_kthreads() 620 int assigned = atomic_read(&ch->kthreads_assigned); in xpc_activate_kthreads() 836 if (atomic_read(&ch->kthreads_assigned) < in xpc_create_kthreads()
|
D | xpc_channel.c | 94 if (atomic_read(&ch->kthreads_assigned) > 0 || in xpc_process_disconnect() 95 atomic_read(&ch->references) > 0) { in xpc_process_disconnect() 123 if (atomic_read(&ch->n_to_notify) > 0) { in xpc_process_disconnect() 136 DBUG_ON(atomic_read(&ch->n_to_notify) != 0); in xpc_process_disconnect() 493 DBUG_ON(atomic_read(&ch->kthreads_assigned) != 0); in xpc_connect_channel() 494 DBUG_ON(atomic_read(&ch->kthreads_idle) != 0); in xpc_connect_channel() 495 DBUG_ON(atomic_read(&ch->kthreads_active) != 0); in xpc_connect_channel() 787 if (atomic_read(&ch->kthreads_idle) > 0) { in xpc_disconnect_channel() 797 if (atomic_read(&ch->n_on_msg_allocate_wq) > 0) in xpc_disconnect_channel()
|
/linux-4.1.27/drivers/usb/misc/ |
D | iowarrior.c | 171 intr_idx = atomic_read(&dev->intr_idx); in iowarrior_callback() 174 read_idx = atomic_read(&dev->read_idx); in iowarrior_callback() 262 read_idx = atomic_read(&dev->read_idx); in read_index() 263 intr_idx = atomic_read(&dev->intr_idx); in read_index() 326 } while (atomic_read(&dev->overflow_flag)); in iowarrior_read() 387 if (atomic_read(&dev->write_busy) == MAX_WRITES_IN_FLIGHT) { in iowarrior_write() 394 (!dev->present || (atomic_read (&dev-> write_busy) < MAX_WRITES_IN_FLIGHT))); in iowarrior_write() 443 retval, atomic_read(&dev->write_busy)); in iowarrior_write() 707 if (atomic_read(&dev->write_busy) < MAX_WRITES_IN_FLIGHT) in iowarrior_poll()
|
/linux-4.1.27/drivers/w1/ |
D | w1_family.c | 89 while (atomic_read(&fent->refcnt)) { in w1_unregister_family() 91 fent->fid, atomic_read(&fent->refcnt)); in w1_unregister_family()
|
/linux-4.1.27/drivers/gpu/drm/mga/ |
D | mga_irq.c | 46 return atomic_read(&dev_priv->vbl_received); in mga_get_vblank_counter() 132 (((cur_fence = atomic_read(&dev_priv->last_fence_retired)) in mga_driver_fence_wait()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | atomic.h | 25 static inline int atomic_read(const atomic_t *v) in atomic_read() function 197 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/drivers/media/platform/s5p-mfc/ |
D | s5p_mfc_pm.c | 104 mfc_debug(3, "+ %d\n", atomic_read(&clk_ref)); in s5p_mfc_clock_on() 114 mfc_debug(3, "- %d\n", atomic_read(&clk_ref)); in s5p_mfc_clock_off()
|
/linux-4.1.27/drivers/crypto/caam/ |
D | caamrng.c | 142 if (atomic_read(&bd->empty)) { in caam_read() 144 if (atomic_read(&bd->empty) == BUF_EMPTY) { in caam_read() 155 if (atomic_read(&bd->empty)) in caam_read() 248 if (atomic_read(&bd->empty) == BUF_PENDING) in caam_cleanup()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/ |
D | l2t.c | 275 if (!atomic_read(&d->nfree)) in alloc_l2e() 280 if (atomic_read(&e->refcnt) == 0) in alloc_l2e() 283 for (e = d->l2tab; atomic_read(&e->refcnt); ++e) in alloc_l2e() 313 if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */ in t4_l2e_free() 388 if (atomic_read(&e->refcnt) == 1) in cxgb4_l2t_get() 490 if (atomic_read(&e->refcnt)) in t4_l2t_update() 633 l2e_state(e), atomic_read(&e->refcnt), in l2t_seq_show()
|
/linux-4.1.27/include/net/ |
D | request_sock.h | 102 WARN_ON_ONCE(atomic_read(&req->rsk_refcnt) != 0); in reqsk_free() 265 return lopt->qlen_inc - atomic_read(&lopt->qlen_dec); in listen_sock_qlen() 270 return lopt->young_inc - atomic_read(&lopt->young_dec); in listen_sock_young()
|
/linux-4.1.27/drivers/target/iscsi/ |
D | iscsi_target_erl0.c | 763 if (atomic_read(&sess->session_reinstatement)) { in iscsit_handle_time2retain_timeout() 851 if (atomic_read(&conn->connection_exit)) { in iscsit_connection_reinstatement_rcfr() 856 if (atomic_read(&conn->transport_failed)) { in iscsit_connection_reinstatement_rcfr() 875 if (atomic_read(&conn->connection_exit)) { in iscsit_cause_connection_reinstatement() 880 if (atomic_read(&conn->transport_failed)) { in iscsit_cause_connection_reinstatement() 885 if (atomic_read(&conn->connection_reinstatement)) { in iscsit_cause_connection_reinstatement() 922 !atomic_read(&sess->session_reinstatement) && in iscsit_handle_connection_cleanup() 923 !atomic_read(&sess->session_fall_back_to_erl0)) in iscsit_handle_connection_cleanup() 936 if (atomic_read(&conn->connection_exit)) { in iscsit_take_action_for_connection_exit()
|
/linux-4.1.27/arch/parisc/kernel/ |
D | ftrace.c | 67 trace->overrun = atomic_read(¤t->trace_overrun); in pop_return_trace() 118 if (unlikely(atomic_read(¤t->tracing_graph_pause))) in prepare_ftrace_return()
|
/linux-4.1.27/fs/ocfs2/ |
D | journal.c | 318 flushed = atomic_read(&journal->j_num_trans); in ocfs2_commit_cache() 336 flushed = atomic_read(&journal->j_num_trans); in ocfs2_commit_cache() 963 num_running_trans = atomic_read(&(osb->journal->j_num_trans)); in ocfs2_journal_shutdown() 982 BUG_ON(atomic_read(&(osb->journal->j_num_trans)) != 0); in ocfs2_journal_shutdown() 1895 if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE) in ocfs2_queue_orphan_scan() 1899 atomic_read(&os->os_state)); in ocfs2_queue_orphan_scan() 1909 if (atomic_read(&os->os_state) == ORPHAN_SCAN_INACTIVE) in ocfs2_queue_orphan_scan() 1931 atomic_read(&os->os_state)); in ocfs2_queue_orphan_scan() 1947 if (atomic_read(&os->os_state) == ORPHAN_SCAN_ACTIVE) in ocfs2_orphan_scan_work() 1958 if (atomic_read(&os->os_state) == ORPHAN_SCAN_ACTIVE) { in ocfs2_orphan_scan_stop() [all …]
|
/linux-4.1.27/security/keys/ |
D | proc.c | 255 atomic_read(&key->usage), in proc_keys_show() 343 atomic_read(&user->usage), in proc_key_users_show() 344 atomic_read(&user->nkeys), in proc_key_users_show() 345 atomic_read(&user->nikeys), in proc_key_users_show()
|
/linux-4.1.27/sound/soc/fsl/ |
D | imx-pcm-fiq.c | 56 if (!atomic_read(&iprtd->playing) && !atomic_read(&iprtd->capturing)) in snd_hrtimer_callback() 137 if (!atomic_read(&iprtd->playing) && in snd_imx_pcm_trigger() 138 !atomic_read(&iprtd->capturing)) in snd_imx_pcm_trigger()
|
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/ |
D | selftest.h | 243 atomic_read(&(rpc)->crpc_refcount)); \ 244 LASSERT(atomic_read(&(rpc)->crpc_refcount) > 0); \ 252 atomic_read(&(rpc)->crpc_refcount)); \ 253 LASSERT(atomic_read(&(rpc)->crpc_refcount) > 0); \ 506 LASSERT (atomic_read(&rpc->crpc_refcount) == 0); in srpc_destroy_client_rpc()
|
/linux-4.1.27/drivers/infiniband/core/ |
D | fmr_pool.c | 182 if (atomic_read(&pool->flush_ser) - atomic_read(&pool->req_ser) < 0) { in ib_fmr_cleanup_thread() 193 if (atomic_read(&pool->flush_ser) - atomic_read(&pool->req_ser) >= 0 && in ib_fmr_cleanup_thread() 418 atomic_read(&pool->flush_ser) - serial >= 0)) in ib_flush_fmr_pool()
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
D | nes_nic.c | 1277 target_stat_values[++index] = atomic_read(&cm_connects); in nes_netdev_get_ethtool_stats() 1278 target_stat_values[++index] = atomic_read(&cm_accepts); in nes_netdev_get_ethtool_stats() 1279 target_stat_values[++index] = atomic_read(&cm_disconnects); in nes_netdev_get_ethtool_stats() 1280 target_stat_values[++index] = atomic_read(&cm_connecteds); in nes_netdev_get_ethtool_stats() 1281 target_stat_values[++index] = atomic_read(&cm_connect_reqs); in nes_netdev_get_ethtool_stats() 1282 target_stat_values[++index] = atomic_read(&cm_rejects); in nes_netdev_get_ethtool_stats() 1283 target_stat_values[++index] = atomic_read(&mod_qp_timouts); in nes_netdev_get_ethtool_stats() 1284 target_stat_values[++index] = atomic_read(&qps_created); in nes_netdev_get_ethtool_stats() 1285 target_stat_values[++index] = atomic_read(&sw_qps_destroyed); in nes_netdev_get_ethtool_stats() 1286 target_stat_values[++index] = atomic_read(&qps_destroyed); in nes_netdev_get_ethtool_stats() [all …]
|
/linux-4.1.27/arch/mn10300/include/asm/ |
D | atomic.h | 37 #define atomic_read(v) (ACCESS_ONCE((v)->counter)) macro 121 c = atomic_read(v); \
|
/linux-4.1.27/drivers/oprofile/ |
D | event_buffer.c | 170 wait_event_interruptible(buffer_wait, atomic_read(&buffer_ready)); in event_buffer_read() 176 if (!atomic_read(&buffer_ready)) in event_buffer_read()
|
/linux-4.1.27/arch/arm/kernel/ |
D | machine_kexec.c | 126 while ((atomic_read(&waiting_for_crash_ipi) > 0) && msecs) { in machine_crash_shutdown() 130 if (atomic_read(&waiting_for_crash_ipi) > 0) in machine_crash_shutdown()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
D | videobuf2-memops.c | 151 __func__, h, atomic_read(h->refcount), vma->vm_start, in vb2_common_vm_open() 169 __func__, h, atomic_read(h->refcount), vma->vm_start, in vb2_common_vm_close()
|
/linux-4.1.27/arch/avr32/include/asm/ |
D | atomic.h | 22 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 104 int tmp, old = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | atomic.h | 24 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 39 old = atomic_read(v); \ 128 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/security/integrity/ima/ |
D | ima_main.c | 91 if (atomic_read(&inode->i_readcount) && IS_IMA(inode)) { in ima_rdwr_violation_check() 99 if ((atomic_read(&inode->i_writecount) > 0) && must_measure) in ima_rdwr_violation_check() 124 if (atomic_read(&inode->i_writecount) == 1) { in ima_check_last_writer()
|
/linux-4.1.27/drivers/target/ |
D | target_core_ua.c | 55 if (!atomic_read(&deve->ua_count)) in target_scsi3_ua_check() 216 if (!atomic_read(&deve->ua_count)) { in core_scsi3_ua_for_check_condition() 284 if (!atomic_read(&deve->ua_count)) { in core_scsi3_ua_clear_for_request_sense()
|
D | target_core_alua.c | 198 buf[off++] |= (atomic_read( in target_emulate_report_target_port_groups() 716 if (atomic_read(&port->sep_tg_pt_secondary_offline)) { in target_alua_state_check() 734 out_alua_state = atomic_read(&tg_pt_gp->tg_pt_gp_alua_access_state); in target_alua_state_check() 1077 if (atomic_read(&tg_pt_gp->tg_pt_gp_alua_access_state) == new_state) in core_alua_do_transition_tg_pt() 1087 atomic_read(&tg_pt_gp->tg_pt_gp_alua_access_state) == in core_alua_do_transition_tg_pt() 1103 atomic_read(&tg_pt_gp->tg_pt_gp_alua_access_state); in core_alua_do_transition_tg_pt() 1285 atomic_read(&port->sep_tg_pt_secondary_offline), in core_alua_update_tpg_secondary_metadata() 1566 while (atomic_read(&lu_gp->lu_gp_ref_cnt)) in core_alua_free_lu_gp() 1614 while (atomic_read(&lu_gp_mem->lu_gp_mem_ref_cnt)) in core_alua_free_lu_gp_mem() 1845 while (atomic_read(&tg_pt_gp->tg_pt_gp_ref_cnt)) in core_alua_free_tg_pt_gp() [all …]
|
/linux-4.1.27/drivers/scsi/ |
D | sg.c | 244 (atomic_read(&sdp->detaching) || in open_wait() 250 if (atomic_read(&sdp->detaching)) in open_wait() 257 (atomic_read(&sdp->detaching) || in open_wait() 263 if (atomic_read(&sdp->detaching)) in open_wait() 446 if (atomic_read(&sdp->detaching)) { in sg_read() 455 (atomic_read(&sdp->detaching) || in sg_read() 457 if (atomic_read(&sdp->detaching)) { in sg_read() 599 if (atomic_read(&sdp->detaching)) in sg_write() 790 if (atomic_read(&sdp->detaching)) { in sg_common_write() 851 if (atomic_read(&sdp->detaching)) in sg_ioctl() [all …]
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
D | svc_rdma_transport.c | 254 if (atomic_read(&xprt->sc_xprt.xpt_ref.refcount)==0) in rq_comp_handler() 435 if (atomic_read(&xprt->sc_xprt.xpt_ref.refcount)==0) in sq_comp_handler() 1134 if (atomic_read(&rdma->sc_xprt.xpt_ref.refcount) != 0) in __svc_rdma_free() 1136 atomic_read(&rdma->sc_xprt.xpt_ref.refcount)); in __svc_rdma_free() 1164 if (atomic_read(&rdma->sc_ctxt_used) != 0) in __svc_rdma_free() 1166 atomic_read(&rdma->sc_ctxt_used)); in __svc_rdma_free() 1167 if (atomic_read(&rdma->sc_dma_used) != 0) in __svc_rdma_free() 1169 atomic_read(&rdma->sc_dma_used)); in __svc_rdma_free() 1276 if (xprt->sc_sq_depth < atomic_read(&xprt->sc_sq_count) + wr_count) { in svc_rdma_send() 1285 atomic_read(&xprt->sc_sq_count) < in svc_rdma_send() [all …]
|
/linux-4.1.27/fs/jbd2/ |
D | commit.c | 72 if (atomic_read(&bh->b_count) != 1) in release_buffer_page() 454 while (atomic_read(&commit_transaction->t_updates)) { in jbd2_journal_commit_transaction() 459 if (atomic_read(&commit_transaction->t_updates)) { in jbd2_journal_commit_transaction() 470 J_ASSERT (atomic_read(&commit_transaction->t_outstanding_credits) <= in jbd2_journal_commit_transaction() 532 atomic_sub(atomic_read(&journal->j_reserved_credits), in jbd2_journal_commit_transaction() 578 atomic_read(&commit_transaction->t_outstanding_credits); in jbd2_journal_commit_transaction() 582 atomic_read(&commit_transaction->t_outstanding_credits)); in jbd2_journal_commit_transaction() 841 J_ASSERT_BH(bh, atomic_read(&bh->b_count) == 0); in jbd2_journal_commit_transaction() 1105 atomic_read(&commit_transaction->t_handle_count); in jbd2_journal_commit_transaction()
|
/linux-4.1.27/drivers/staging/android/ |
D | sync.c | 330 int err = atomic_read(&fence->status); in sync_fence_wait_async() 343 err = atomic_read(&fence->status); in sync_fence_wait_async() 385 atomic_read(&fence->status) <= 0, in sync_fence_wait() 400 ret = atomic_read(&fence->status); in sync_fence_wait() 522 int i, status = atomic_read(&fence->status); in sync_fence_free() 551 status = atomic_read(&fence->status); in sync_fence_poll() 673 data->status = atomic_read(&fence->status); in sync_fence_ioctl_fence_info()
|
/linux-4.1.27/security/tomoyo/ |
D | gc.c | 439 if (atomic_read(&container_of in tomoyo_try_to_gc() 522 if (!domain->is_deleted || atomic_read(&domain->users)) in tomoyo_collect_entry() 538 if (atomic_read(&ptr->users) > 0) in tomoyo_collect_entry() 563 atomic_read(&group->head.users) > 0) in tomoyo_collect_entry() 577 if (atomic_read(&ptr->users) > 0) in tomoyo_collect_entry()
|
/linux-4.1.27/fs/nfsd/ |
D | nfscache.c | 251 if (atomic_read(&num_drc_entries) <= max_drc_entries && in prune_bucket() 301 return atomic_read(&num_drc_entries); in nfsd_reply_cache_count() 390 longest_chain_cachesize = atomic_read(&num_drc_entries); in nfsd_cache_search() 395 atomic_read(&num_drc_entries)); in nfsd_cache_search() 622 atomic_read(&num_drc_entries)); in nfsd_reply_cache_stats_show()
|
/linux-4.1.27/arch/arc/include/asm/ |
D | atomic.h | 20 #define atomic_read(v) ((v)->counter) macro 163 c = atomic_read(v); \
|
/linux-4.1.27/arch/powerpc/mm/ |
D | icswx.c | 113 if (atomic_read(&mm->mm_users) > 1) in use_cop() 153 if (atomic_read(&mm->mm_users) > 1) in drop_cop()
|
/linux-4.1.27/arch/arm/mach-exynos/ |
D | pm.c | 216 if (atomic_read(&cpu1_wakeup)) in exynos_cpu0_enter_aftr() 250 while (!atomic_read(&cpu1_wakeup)) { in exynos_cpu0_enter_aftr()
|
/linux-4.1.27/drivers/s390/net/ |
D | fsm.h | 146 int state = atomic_read(&fi->state); in fsm_event() 214 return atomic_read(&fi->state); in fsm_getstate()
|
/linux-4.1.27/drivers/s390/cio/ |
D | crw.c | 69 atomic_read(&crw_nr_req) > 0); in crw_collect_info() 144 wait_event(crw_handler_wait_q, atomic_read(&crw_nr_req) == 0); in crw_wait_for_channel_report()
|
/linux-4.1.27/fs/proc/ |
D | task_nommu.c | 39 if (atomic_read(&mm->mm_count) > 1 || in task_mem() 49 if (atomic_read(&mm->mm_count) > 1) in task_mem() 59 if (current->files && atomic_read(¤t->files->count) > 1) in task_mem() 64 if (current->sighand && atomic_read(¤t->sighand->count) > 1) in task_mem()
|
/linux-4.1.27/drivers/media/usb/tm6000/ |
D | tm6000-alsa.c | 201 if (atomic_read(&core->stream_started) > 0) { in snd_tm6000_close() 218 if (atomic_read(&core->stream_started) == 0) in tm6000_fillbuf() 304 if (atomic_read(&core->stream_started) > 0) { in snd_tm6000_hw_free() 336 if (atomic_read(&core->stream_started)) { in audio_trigger()
|
/linux-4.1.27/drivers/misc/sgi-gru/ |
D | grutlbpurge.c | 232 start, end, atomic_read(&gms->ms_range_active)); in gru_invalidate_range_start() 321 atomic_read(&gms->ms_refcnt)); in gru_register_mmu_notifier() 331 atomic_read(&gms->ms_refcnt), gms->ms_released); in gru_drop_mmu_notifier()
|
/linux-4.1.27/kernel/rcu/ |
D | rcutorture.c | 1024 if (atomic_read(&beenhere)) in rcutorture_trace_dump() 1208 atomic_read(&n_rcu_torture_alloc), in rcu_torture_stats_print() 1209 atomic_read(&n_rcu_torture_alloc_fail), in rcu_torture_stats_print() 1210 atomic_read(&n_rcu_torture_free)); in rcu_torture_stats_print() 1212 atomic_read(&n_rcu_torture_mberror), in rcu_torture_stats_print() 1227 if (atomic_read(&n_rcu_torture_mberror) != 0 || in rcu_torture_stats_print() 1251 pr_cont(" %d", atomic_read(&rcu_torture_wcount[i])); in rcu_torture_stats_print() 1447 atomic_read(&barrier_cbs_count) == 0 || in rcu_torture_barrier() 1453 if (atomic_read(&barrier_cbs_invoked) != n_barrier_cbs) { in rcu_torture_barrier() 1456 atomic_read(&barrier_cbs_invoked), in rcu_torture_barrier() [all …]
|
/linux-4.1.27/sound/core/ |
D | hrtimer.c | 50 if (!atomic_read(&stime->running)) in snd_hrtimer_callback() 56 if (!atomic_read(&stime->running)) in snd_hrtimer_callback()
|
/linux-4.1.27/arch/xtensa/include/asm/ |
D | atomic.h | 50 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 241 c = atomic_read(v); in ATOMIC_OPS()
|
/linux-4.1.27/fs/configfs/ |
D | configfs_internal.h | 151 WARN_ON(!atomic_read(&sd->s_count)); in configfs_get() 159 WARN_ON(!atomic_read(&sd->s_count)); in configfs_put()
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | atomic_64.h | 17 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 78 c = atomic_read(v); in ATOMIC_OPS()
|
/linux-4.1.27/net/rds/ |
D | ib_rdma.c | 258 WARN_ON(atomic_read(&pool->item_count)); in rds_ib_destroy_mr_pool() 259 WARN_ON(atomic_read(&pool->free_pinned)); in rds_ib_destroy_mr_pool() 299 if (atomic_read(&pool->dirty_count) >= pool->max_items / 10) in rds_ib_alloc_fmr() 516 item_count = atomic_read(&pool->item_count); in rds_ib_flush_goal() 712 if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned || in rds_ib_free_mr() 713 atomic_read(&pool->dirty_count) >= pool->max_items / 10) in rds_ib_free_mr()
|
/linux-4.1.27/fs/kernfs/ |
D | dir.c | 30 return atomic_read(&kn->active) >= 0; in kernfs_active() 356 if (atomic_read(&kn->active) != KN_DEACTIVATED_BIAS) in kernfs_drain() 362 atomic_read(&kn->active) == KN_DEACTIVATED_BIAS); in kernfs_drain() 381 WARN_ON(!atomic_read(&kn->count)); in kernfs_get() 408 WARN_ONCE(atomic_read(&kn->active) != KN_DEACTIVATED_BIAS, in kernfs_put() 410 parent ? parent->name : "", kn->name, atomic_read(&kn->active)); in kernfs_put() 1018 WARN_ON_ONCE(atomic_read(&pos->active) != KN_DEACTIVATED_BIAS); in kernfs_activate() 1070 WARN_ON_ONCE(atomic_read(&kn->active) != KN_DEACTIVATED_BIAS); in __kernfs_remove() 1215 atomic_read(&kn->active) == KN_DEACTIVATED_BIAS) in kernfs_remove_self()
|
/linux-4.1.27/arch/m32r/include/asm/ |
D | atomic.h | 31 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 230 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/arch/frv/include/asm/ |
D | atomic.h | 34 #define atomic_read(v) ACCESS_ONCE((v)->counter) macro 187 c = atomic_read(v); in __atomic_add_unless()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | nmi.c | 151 if (!atomic_read(&nmi_touched[this_cpu])) in check_nmi_wdt_touched() 160 if (!atomic_read(&nmi_touched[cpu])) in check_nmi_wdt_touched()
|
/linux-4.1.27/drivers/misc/ibmasm/ |
D | command.c | 61 dbg("command count: %d\n", atomic_read(&command_count)); in ibmasm_new_command() 72 dbg("command count: %d\n", atomic_read(&command_count)); in ibmasm_free_command()
|
/linux-4.1.27/kernel/debug/kdb/ |
D | kdb_debugger.c | 39 kdb_initial_cpu = atomic_read(&kgdb_active); in kdb_common_init_state() 69 if (atomic_read(&kgdb_setting_breakpoint)) in kdb_stub()
|
/linux-4.1.27/fs/f2fs/ |
D | debug.c | 39 si->ext_node = atomic_read(&sbi->total_ext_node); in update_general_status() 52 si->inline_inode = atomic_read(&sbi->inline_inode); in update_general_status() 53 si->inline_dir = atomic_read(&sbi->inline_dir); in update_general_status() 88 si->inplace_count = atomic_read(&sbi->inplace_count); in update_general_status() 191 si->cache_mem += atomic_read(&sbi->total_ext_node) * in update_mem_info()
|
/linux-4.1.27/drivers/staging/lustre/lustre/llite/ |
D | llite_capa.c | 143 return atomic_read(&lli->lli_open_count); in obd_capa_open_count() 265 if (atomic_read(&ocapa->c_refc) > 1) { in capa_thread_main() 268 atomic_read(&ocapa->c_refc)); in capa_thread_main() 361 if (atomic_read(&ll_capa_debug)) { in ll_osscapa_get() 386 if (!ocapa && atomic_read(&ll_capa_debug)) { in ll_mdscapa_get()
|
/linux-4.1.27/arch/m68k/kernel/ |
D | irq.c | 37 seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); in arch_show_interrupts()
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_irq.c | 137 tslot = atomic_read(&vblank->count) + diff; in drm_update_vblank_count() 220 vblcount = atomic_read(&vblank->count); in vblank_disable_and_save() 262 if (atomic_read(&vblank->refcount) == 0 && vblank->enabled) { in vblank_disable_fn() 854 return atomic_read(&vblank->count); in drm_vblank_count() 905 cur_vblank = atomic_read(&vblank->count); in drm_vblank_count_and_time() 908 } while (cur_vblank != atomic_read(&vblank->count)); in drm_vblank_count_and_time() 1083 if (WARN_ON(atomic_read(&vblank->refcount) == 0)) in drm_vblank_put() 1311 if (atomic_read(&vblank->refcount) != 0 || in drm_vblank_on() 1717 vblcount = atomic_read(&vblank->count); in drm_handle_vblank()
|
/linux-4.1.27/include/misc/ |
D | cxl.h | 26 return (atomic_read(&cxl_use_count) != 0); in cxl_ctx_in_use()
|
/linux-4.1.27/net/bluetooth/hidp/ |
D | core.c | 106 if (atomic_read(&session->terminate)) in hidp_send_message() 238 if (atomic_read(&session->terminate)) in hidp_get_raw_report() 270 !atomic_read(&session->terminate)) { in hidp_get_raw_report() 275 || atomic_read(&session->terminate), in hidp_get_raw_report() 345 !atomic_read(&session->terminate)) { in hidp_set_raw_report() 350 || atomic_read(&session->terminate), in hidp_set_raw_report() 1058 while (atomic_read(&session->state) <= HIDP_SESSION_IDLING) in hidp_session_start_sync() 1060 atomic_read(&session->state) > HIDP_SESSION_IDLING); in hidp_session_start_sync() 1161 atomic_read(&session->state) > HIDP_SESSION_PREPARING) in hidp_session_remove() 1196 if (atomic_read(&session->terminate)) in hidp_session_run()
|
/linux-4.1.27/drivers/message/fusion/ |
D | mptlan.c | 507 priv->total_posted,atomic_read(&priv->buckets_out))); in mpt_lan_close() 514 while (atomic_read(&priv->buckets_out) && time_before(jiffies, timeout)) in mpt_lan_close() 844 atomic_read(&priv->buckets_out))); in mpt_lan_receive_skb() 846 if (atomic_read(&priv->buckets_out) < priv->bucketthresh) in mpt_lan_receive_skb() 851 atomic_read(&priv->buckets_out), priv->total_received)); in mpt_lan_receive_skb() 962 /**/ atomic_read(&priv->buckets_out), priv->total_received)); in mpt_lan_receive_post_free() 1119 atomic_read(&priv->buckets_out)); in mpt_lan_receive_post_reply() 1124 remaining, atomic_read(&priv->buckets_out)); in mpt_lan_receive_post_reply() 1127 ((atomic_read(&priv->buckets_out) - remaining) > in mpt_lan_receive_post_reply() 1163 curr = atomic_read(&priv->buckets_out); in mpt_lan_post_receive_buckets() [all …]
|
/linux-4.1.27/net/sunrpc/ |
D | bc_svc.c | 56 WARN_ON_ONCE(atomic_read(&task->tk_count) != 1); in bc_send()
|