/linux-4.1.27/drivers/connector/ |
H A D | cn_queue.c | 85 spin_lock_bh(&dev->queue_lock); cn_queue_add_callback() 94 spin_unlock_bh(&dev->queue_lock); cn_queue_add_callback() 112 spin_lock_bh(&dev->queue_lock); cn_queue_del_callback() 120 spin_unlock_bh(&dev->queue_lock); cn_queue_del_callback() 137 spin_lock_init(&dev->queue_lock); cn_queue_alloc_dev() 148 spin_lock_bh(&dev->queue_lock); cn_queue_free_dev() 151 spin_unlock_bh(&dev->queue_lock); cn_queue_free_dev()
|
H A D | connector.c | 88 spin_lock_bh(&dev->cbdev->queue_lock); cn_netlink_send_mult() 97 spin_unlock_bh(&dev->cbdev->queue_lock); cn_netlink_send_mult() 156 spin_lock_bh(&dev->cbdev->queue_lock); cn_call_callback() 164 spin_unlock_bh(&dev->cbdev->queue_lock); cn_call_callback() 248 spin_lock_bh(&dev->queue_lock); cn_proc_show() 257 spin_unlock_bh(&dev->queue_lock); cn_proc_show()
|
/linux-4.1.27/drivers/scsi/arm/ |
H A D | queue.c | 63 spin_lock_init(&queue->queue_lock); queue_initialise() 113 spin_lock_irqsave(&queue->queue_lock, flags); __queue_add() 133 spin_unlock_irqrestore(&queue->queue_lock, flags); __queue_add() 167 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_exclude() 176 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_exclude() 192 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove() 195 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove() 216 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_tgtluntag() 225 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_tgtluntag() 242 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_all_target() 248 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_all_target() 266 spin_lock_irqsave(&queue->queue_lock, flags); queue_probetgtlun() 274 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_probetgtlun() 292 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_cmd() 301 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_cmd()
|
H A D | queue.h | 16 spinlock_t queue_lock; member in struct:__anon8658
|
/linux-4.1.27/drivers/isdn/i4l/ |
H A D | isdn_net.h | 84 spin_lock_irqsave(&nd->queue_lock, flags); isdn_net_get_locked_lp() 95 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_get_locked_lp() 100 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_get_locked_lp() 112 spin_lock_irqsave(&nd->queue_lock, flags); isdn_net_add_to_bundle() 123 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_add_to_bundle() 138 spin_lock_irqsave(&master_lp->netdev->queue_lock, flags); isdn_net_rm_from_bundle() 150 spin_unlock_irqrestore(&master_lp->netdev->queue_lock, flags); isdn_net_rm_from_bundle()
|
H A D | isdn_net.c | 128 spin_lock_irqsave(&nd->queue_lock, flags); isdn_net_device_busy() 132 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_device_busy() 137 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_device_busy() 2593 spin_lock_init(&netdev->queue_lock); isdn_net_new()
|
/linux-4.1.27/block/ |
H A D | blk-ioc.c | 60 lockdep_assert_held(q->queue_lock); ioc_destroy_icq() 68 * under queue_lock. If it's not pointing to @icq now, it never ioc_destroy_icq() 107 if (spin_trylock(q->queue_lock)) { ioc_release_fn() 109 spin_unlock(q->queue_lock); ioc_release_fn() 141 * already be holding a queue_lock. Do it asynchronously from wq. put_io_context() 185 if (spin_trylock(icq->q->queue_lock)) { put_io_context_active() 187 spin_unlock(icq->q->queue_lock); put_io_context_active() 221 lockdep_assert_held(q->queue_lock); ioc_clear_queue() 315 * with @q->queue_lock held. 321 lockdep_assert_held(q->queue_lock); ioc_lookup_icq() 380 spin_lock_irq(q->queue_lock); ioc_create_icq() 396 spin_unlock_irq(q->queue_lock); ioc_create_icq()
|
H A D | blk-cgroup.c | 125 * holding @q->queue_lock and lookup hint is updated on success. 138 * hint can only be updated under queue_lock as otherwise @blkg __blkg_lookup() 140 * responsible for grabbing queue_lock if @update_hint. __blkg_lookup() 145 lockdep_assert_held(q->queue_lock); __blkg_lookup() 185 lockdep_assert_held(q->queue_lock); blkg_create() 265 * should be called under RCU read lock and @q->queue_lock. 277 lockdep_assert_held(q->queue_lock); blkg_lookup_create() 315 lockdep_assert_held(blkg->q->queue_lock); blkg_destroy() 336 * under queue_lock. If it's not pointing to @blkg now, it never blkg_destroy() 368 lockdep_assert_held(q->queue_lock); blkg_destroy_all() 524 spin_lock_irq(blkg->q->queue_lock); blkcg_print_blkgs() 527 spin_unlock_irq(blkg->q->queue_lock); blkcg_print_blkgs() 637 lockdep_assert_held(pd->blkg->q->queue_lock); blkg_stat_recursive_sum() 671 lockdep_assert_held(pd->blkg->q->queue_lock); blkg_rwstat_recursive_sum() 707 __acquires(rcu) __acquires(disk->queue->queue_lock) __acquires() 727 spin_lock_irq(disk->queue->queue_lock); __acquires() 737 spin_unlock_irq(disk->queue->queue_lock); __acquires() 767 __releases(ctx->disk->queue->queue_lock) __releases(rcu) __releases() 769 spin_unlock_irq(ctx->disk->queue->queue_lock); __releases() 805 if (spin_trylock(q->queue_lock)) { blkcg_css_offline() 807 spin_unlock(q->queue_lock); blkcg_css_offline() 875 lockdep_assert_held(q->queue_lock); blkcg_drain_queue() 895 spin_lock_irq(q->queue_lock); blkcg_exit_queue() 897 spin_unlock_irq(q->queue_lock); blkcg_exit_queue() 987 spin_lock_irq(q->queue_lock); blkcg_activate_policy() 1008 spin_unlock_irq(q->queue_lock); blkcg_activate_policy() 1024 spin_lock_irq(q->queue_lock); blkcg_activate_policy() 1049 spin_unlock_irq(q->queue_lock); blkcg_activate_policy() 1075 spin_lock_irq(q->queue_lock); blkcg_deactivate_policy() 1098 spin_unlock_irq(q->queue_lock); blkcg_deactivate_policy()
|
H A D | blk-exec.c | 79 spin_lock_irq(q->queue_lock); blk_execute_rq_nowait() 85 spin_unlock_irq(q->queue_lock); blk_execute_rq_nowait() 94 spin_unlock_irq(q->queue_lock); blk_execute_rq_nowait()
|
H A D | blk-core.c | 163 spin_lock_irq(q->queue_lock); blk_delay_work() 165 spin_unlock_irq(q->queue_lock); blk_delay_work() 333 spin_lock_irqsave(q->queue_lock, flags); blk_run_queue() 335 spin_unlock_irqrestore(q->queue_lock, flags); blk_run_queue() 355 __releases(q->queue_lock) 356 __acquires(q->queue_lock) 360 lockdep_assert_held(q->queue_lock); 406 spin_unlock_irq(q->queue_lock); 410 spin_lock_irq(q->queue_lock); 439 spin_lock_irq(q->queue_lock); blk_queue_bypass_start() 442 spin_unlock_irq(q->queue_lock); blk_queue_bypass_start() 450 spin_lock_irq(q->queue_lock); blk_queue_bypass_start() 452 spin_unlock_irq(q->queue_lock); blk_queue_bypass_start() 468 spin_lock_irq(q->queue_lock); blk_queue_bypass_end() 472 spin_unlock_irq(q->queue_lock); blk_queue_bypass_end() 504 spinlock_t *lock = q->queue_lock; blk_cleanup_queue() 551 if (q->queue_lock != &q->__queue_lock) blk_cleanup_queue() 552 q->queue_lock = &q->__queue_lock; blk_cleanup_queue() 650 * By default initialize queue_lock to internal lock and driver can blk_alloc_queue_node() 653 q->queue_lock = &q->__queue_lock; blk_alloc_queue_node() 760 q->queue_lock = lock; blk_init_allocated_queue() 866 * congestion status, wake up any waiters. Called under q->queue_lock. 888 spin_lock_irq(q->queue_lock); blk_update_nr_requests() 921 spin_unlock_irq(q->queue_lock); 970 * Must be called with @q->queue_lock held and, 971 * Returns ERR_PTR on failure, with @q->queue_lock held. 972 * Returns request pointer on success, with @q->queue_lock *not held*. 1042 * Also, lookup icq while holding queue_lock. If it doesn't exist, __get_request() 1043 * it will be created after releasing queue_lock. __get_request() 1054 spin_unlock_irq(q->queue_lock); __get_request() 1108 spin_lock_irq(q->queue_lock); __get_request() 1110 spin_unlock_irq(q->queue_lock); __get_request() 1121 spin_lock_irq(q->queue_lock); __get_request() 1147 * Must be called with @q->queue_lock held and, 1148 * Returns ERR_PTR on failure, with @q->queue_lock held. 1149 * Returns request pointer on success, with @q->queue_lock *not held*. 1176 spin_unlock_irq(q->queue_lock); get_request() 1186 spin_lock_irq(q->queue_lock); get_request() 1202 spin_lock_irq(q->queue_lock); blk_old_get_request() 1205 spin_unlock_irq(q->queue_lock); blk_old_get_request() 1206 /* q->queue_lock is unlocked at this point */ blk_old_get_request() 1424 spin_lock_irqsave(q->queue_lock, flags); blk_put_request() 1426 spin_unlock_irqrestore(q->queue_lock, flags); blk_put_request() 1519 * going through @q->queue_lock. As such it's more of an issuing mechanism 1604 spin_lock_irq(q->queue_lock); blk_queue_bio() 1617 spin_lock_irq(q->queue_lock); blk_queue_bio() 1684 spin_lock_irq(q->queue_lock); blk_queue_bio() 1688 spin_unlock_irq(q->queue_lock); blk_queue_bio() 2074 spin_lock_irqsave(q->queue_lock, flags); blk_insert_cloned_request() 2076 spin_unlock_irqrestore(q->queue_lock, flags); blk_insert_cloned_request() 2092 spin_unlock_irqrestore(q->queue_lock, flags); blk_insert_cloned_request() 2112 * queue_lock must be held. 2254 * queue_lock must be held. 2377 * queue_lock must be held. 2409 * queue_lock must be held. 2654 spin_lock_irqsave(q->queue_lock, flags); blk_end_bidi_request() 2656 spin_unlock_irqrestore(q->queue_lock, flags); blk_end_bidi_request() 3072 __releases(q->queue_lock) 3080 spin_unlock(q->queue_lock); 3165 spin_lock(q->queue_lock); blk_flush_plug_list() 3261 spin_lock_irq(q->queue_lock); blk_pre_runtime_suspend() 3268 spin_unlock_irq(q->queue_lock); blk_pre_runtime_suspend() 3288 spin_lock_irq(q->queue_lock); blk_post_runtime_suspend() 3295 spin_unlock_irq(q->queue_lock); blk_post_runtime_suspend() 3312 spin_lock_irq(q->queue_lock); blk_pre_runtime_resume() 3314 spin_unlock_irq(q->queue_lock); blk_pre_runtime_resume() 3334 spin_lock_irq(q->queue_lock); blk_post_runtime_resume() 3343 spin_unlock_irq(q->queue_lock); blk_post_runtime_resume()
|
H A D | blk-timeout.c | 60 spin_lock_irq(q->queue_lock); part_timeout_store() 65 spin_unlock_irq(q->queue_lock); part_timeout_store() 137 spin_lock_irqsave(q->queue_lock, flags); blk_rq_timed_out_timer() 145 spin_unlock_irqrestore(q->queue_lock, flags); blk_rq_timed_out_timer()
|
H A D | noop-iosched.c | 80 spin_lock_irq(q->queue_lock); noop_init_queue() 82 spin_unlock_irq(q->queue_lock); noop_init_queue()
|
H A D | blk-sysfs.c | 179 spin_lock_irq(q->queue_lock); queue_max_sectors_store() 181 spin_unlock_irq(q->queue_lock); queue_max_sectors_store() 212 spin_lock_irq(q->queue_lock); \ 217 spin_unlock_irq(q->queue_lock); \ 241 spin_lock_irq(q->queue_lock); queue_nomerges_store() 248 spin_unlock_irq(q->queue_lock); queue_nomerges_store() 272 spin_lock_irq(q->queue_lock); queue_rq_affinity_store() 283 spin_unlock_irq(q->queue_lock); queue_rq_affinity_store() 507 spin_lock_irq(q->queue_lock); blk_release_queue() 509 spin_unlock_irq(q->queue_lock); blk_release_queue()
|
H A D | bsg-lib.c | 177 spin_unlock_irq(q->queue_lock); bsg_request_fn() 183 spin_lock_irq(q->queue_lock); bsg_request_fn() 189 spin_lock_irq(q->queue_lock); bsg_request_fn() 194 spin_unlock_irq(q->queue_lock); bsg_request_fn() 196 spin_lock_irq(q->queue_lock); bsg_request_fn()
|
H A D | elevator.c | 582 lockdep_assert_held(q->queue_lock); elv_drain_elevator() 678 spin_lock_irqsave(q->queue_lock, flags); elv_add_request() 680 spin_unlock_irqrestore(q->queue_lock, flags); elv_add_request() 911 spin_lock_irq(q->queue_lock); elevator_switch() 913 spin_unlock_irq(q->queue_lock); elevator_switch()
|
H A D | blk-throttle.c | 1206 spin_lock_irq(q->queue_lock); throtl_pending_timer_fn() 1226 spin_unlock_irq(q->queue_lock); throtl_pending_timer_fn() 1228 spin_lock_irq(q->queue_lock); throtl_pending_timer_fn() 1250 spin_unlock_irq(q->queue_lock); throtl_pending_timer_fn() 1274 spin_lock_irq(q->queue_lock); blk_throtl_dispatch_work_fn() 1278 spin_unlock_irq(q->queue_lock); blk_throtl_dispatch_work_fn() 1394 * We're already holding queue_lock and know @tg is valid. Let's tg_set_conf() 1514 spin_lock_irq(q->queue_lock); blk_throtl_bio() 1582 spin_unlock_irq(q->queue_lock); blk_throtl_bio() 1625 __releases(q->queue_lock) __acquires(q->queue_lock) 1649 spin_unlock_irq(q->queue_lock); 1657 spin_lock_irq(q->queue_lock);
|
H A D | blk-flush.c | 160 * spin_lock_irq(q->queue_lock or fq->mq_flush_lock) 279 * spin_lock_irq(q->queue_lock or fq->mq_flush_lock) 369 * spin_lock_irq(q->queue_lock) in !mq case
|
H A D | deadline-iosched.c | 366 spin_lock_irq(q->queue_lock); deadline_init_queue() 368 spin_unlock_irq(q->queue_lock); deadline_init_queue()
|
H A D | cfq-iosched.c | 487 /* This should be called with the queue_lock held. */ cfqg_stats_update_group_wait_time() 502 /* This should be called with the queue_lock held. */ cfqg_stats_set_start_group_wait_time() 516 /* This should be called with the queue_lock held. */ cfqg_stats_end_empty_time() 743 lockdep_assert_held(cfqg_to_blkg(cfqg)->q->queue_lock); cfqg_stats_xfer_dead() 2845 * Must be called with the queue_lock held. 3611 spin_unlock_irq(cfqd->queue->queue_lock); cfq_find_alloc_queue() 3615 spin_lock_irq(cfqd->queue->queue_lock); cfq_find_alloc_queue() 4226 spin_lock_irq(q->queue_lock); cfq_set_request() 4262 spin_unlock_irq(q->queue_lock); cfq_set_request() 4272 spin_lock_irq(q->queue_lock); cfq_kick_queue() 4274 spin_unlock_irq(q->queue_lock); cfq_kick_queue() 4289 spin_lock_irqsave(cfqd->queue->queue_lock, flags); cfq_idle_slice_timer() 4330 spin_unlock_irqrestore(cfqd->queue->queue_lock, flags); cfq_idle_slice_timer() 4361 spin_lock_irq(q->queue_lock); cfq_exit_queue() 4368 spin_unlock_irq(q->queue_lock); cfq_exit_queue() 4399 spin_lock_irq(q->queue_lock); cfq_init_queue() 4401 spin_unlock_irq(q->queue_lock); cfq_init_queue() 4443 spin_lock_irq(q->queue_lock); cfq_init_queue() 4446 spin_unlock_irq(q->queue_lock); cfq_init_queue()
|
H A D | blk-cgroup.h | 321 * under queue_lock. This function is guaranteed to return non-%NULL 360 * queue_lock. 398 * Should be used under queue_lock.
|
H A D | blk-mq.c | 117 spin_lock_irq(q->queue_lock); blk_mq_freeze_queue_start() 119 spin_unlock_irq(q->queue_lock); blk_mq_freeze_queue_start() 148 spin_lock_irq(q->queue_lock); blk_mq_unfreeze_queue() 151 spin_unlock_irq(q->queue_lock); blk_mq_unfreeze_queue()
|
/linux-4.1.27/drivers/scsi/device_handler/ |
H A D | scsi_dh.c | 126 spin_lock_irq(sdev->request_queue->queue_lock); scsi_dh_handler_attach() 128 spin_unlock_irq(sdev->request_queue->queue_lock); scsi_dh_handler_attach() 141 spin_lock_irq(sdev->request_queue->queue_lock); __detach_handler() 143 spin_unlock_irq(sdev->request_queue->queue_lock); __detach_handler() 418 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_activate() 421 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_activate() 437 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_activate() 470 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_set_params() 476 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_set_params() 514 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_attach() 518 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_attach() 543 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_detach() 547 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_detach() 575 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_attached_handler_name() 579 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_attached_handler_name()
|
/linux-4.1.27/drivers/block/ |
H A D | nbd.c | 48 spinlock_t queue_lock; member in struct:nbd_device 108 spin_lock_irqsave(q->queue_lock, flags); nbd_end_request() 110 spin_unlock_irqrestore(q->queue_lock, flags); nbd_end_request() 290 spin_lock(&nbd->queue_lock); nbd_find_request() 295 spin_unlock(&nbd->queue_lock); nbd_find_request() 298 spin_unlock(&nbd->queue_lock); nbd_find_request() 491 spin_lock(&nbd->queue_lock); nbd_handle_req() 493 spin_unlock(&nbd->queue_lock); nbd_handle_req() 523 spin_lock_irq(&nbd->queue_lock); nbd_thread() 527 spin_unlock_irq(&nbd->queue_lock); nbd_thread() 543 __releases(q->queue_lock) __acquires(q->queue_lock) 550 spin_unlock_irq(q->queue_lock); 564 spin_lock_irq(q->queue_lock); 568 spin_lock_irq(&nbd->queue_lock); 570 spin_unlock_irq(&nbd->queue_lock); 574 spin_lock_irq(q->queue_lock); 844 spin_lock_init(&nbd_dev[i].queue_lock); nbd_init()
|
H A D | DAC960.c | 528 spin_unlock_irq(&Controller->queue_lock); DAC960_WaitForCommand() 530 spin_lock_irq(&Controller->queue_lock); DAC960_WaitForCommand() 790 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_ExecuteCommand() 792 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_ExecuteCommand() 2043 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration() 2045 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration() 2088 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration() 2090 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration() 2536 RequestQueue = blk_init_queue(DAC960_RequestFunction,&Controller->queue_lock); DAC960_RegisterBlockDevice() 2780 spin_lock_init(&Controller->queue_lock); DAC960_DetectController() 3119 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_FinalizeController() 3121 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_FinalizeController() 5271 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_GEM_InterruptHandler() 5295 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_GEM_InterruptHandler() 5312 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_BA_InterruptHandler() 5336 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_BA_InterruptHandler() 5354 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_LP_InterruptHandler() 5378 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_LP_InterruptHandler() 5396 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_LA_InterruptHandler() 5416 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_LA_InterruptHandler() 5434 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_PG_InterruptHandler() 5454 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_PG_InterruptHandler() 5471 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_PD_InterruptHandler() 5488 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_PD_InterruptHandler() 5509 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_P_InterruptHandler() 5561 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_P_InterruptHandler() 5628 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction() 5636 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction() 5686 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction() 5694 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction() 5971 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand() 5974 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand() 6165 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand() 6167 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand() 6234 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand() 6237 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand() 6428 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand() 6430 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand() 6765 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6771 spin_unlock_irq(&Controller->queue_lock); DAC960_gam_ioctl() 6775 spin_lock_irq(&Controller->queue_lock); DAC960_gam_ioctl() 6779 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6790 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6793 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6804 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6806 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6911 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6914 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6962 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl() 6964 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl()
|
H A D | hd.c | 540 spin_lock_irq(hd_queue->queue_lock); hd_times_out() 551 spin_unlock_irq(hd_queue->queue_lock); hd_times_out() 680 spin_lock(hd_queue->queue_lock); hd_interrupt() 688 spin_unlock(hd_queue->queue_lock); hd_interrupt()
|
H A D | null_blk.c | 348 spin_unlock_irq(q->queue_lock); null_request_fn() 350 spin_lock_irq(q->queue_lock); null_request_fn()
|
H A D | nvme-core.c | 604 spin_lock_irqsave(req->q->queue_lock, flags); req_completion() 607 spin_unlock_irqrestore(req->q->queue_lock, flags); req_completion() 2593 spin_lock(ns->queue->queue_lock); nvme_freeze_queues() 2595 spin_unlock(ns->queue->queue_lock); nvme_freeze_queues()
|
H A D | floppy.c | 2251 spin_lock_irqsave(q->queue_lock, flags); request_done() 2253 spin_unlock_irqrestore(q->queue_lock, flags); request_done() 2265 spin_lock_irqsave(q->queue_lock, flags); request_done() 2267 spin_unlock_irqrestore(q->queue_lock, flags); request_done()
|
H A D | umem.c | 898 card->queue->queue_lock = &card->lock; mm_pci_probe()
|
H A D | pktcdvd.c | 2220 spin_lock_irq(q->queue_lock); pkt_open_dev() 2222 spin_unlock_irq(q->queue_lock); pkt_open_dev()
|
H A D | cciss.c | 3314 spin_unlock_irq(q->queue_lock); do_cciss_request() 3428 spin_lock_irq(q->queue_lock); do_cciss_request()
|
H A D | DAC960.h | 2325 spinlock_t queue_lock; member in struct:DAC960_Controller
|
/linux-4.1.27/drivers/mtd/ |
H A D | mtd_blkdevs.c | 137 spin_lock_irq(rq->queue_lock); mtd_blktrans_work() 145 spin_unlock_irq(rq->queue_lock); mtd_blktrans_work() 149 spin_lock_irq(rq->queue_lock); mtd_blktrans_work() 160 spin_unlock_irq(rq->queue_lock); mtd_blktrans_work() 166 spin_lock_irq(rq->queue_lock); mtd_blktrans_work() 174 spin_unlock_irq(rq->queue_lock); mtd_blktrans_work() 409 spin_lock_init(&new->queue_lock); add_mtd_blktrans_dev() 410 new->rq = blk_init_queue(mtd_blktrans_request, &new->queue_lock); add_mtd_blktrans_dev() 481 spin_lock_irqsave(&old->queue_lock, flags); del_mtd_blktrans_dev() 484 spin_unlock_irqrestore(&old->queue_lock, flags); del_mtd_blktrans_dev()
|
/linux-4.1.27/drivers/media/platform/omap3isp/ |
H A D | ispvideo.c | 572 mutex_lock(&video->queue_lock); omap3isp_video_resume() 574 mutex_unlock(&video->queue_lock); omap3isp_video_resume() 827 mutex_lock(&video->queue_lock); isp_video_reqbufs() 829 mutex_unlock(&video->queue_lock); isp_video_reqbufs() 841 mutex_lock(&video->queue_lock); isp_video_querybuf() 843 mutex_unlock(&video->queue_lock); isp_video_querybuf() 855 mutex_lock(&video->queue_lock); isp_video_qbuf() 857 mutex_unlock(&video->queue_lock); isp_video_qbuf() 869 mutex_lock(&video->queue_lock); isp_video_dqbuf() 871 mutex_unlock(&video->queue_lock); isp_video_dqbuf() 1072 mutex_lock(&video->queue_lock); isp_video_streamon() 1074 mutex_unlock(&video->queue_lock); isp_video_streamon() 1097 mutex_lock(&video->queue_lock); isp_video_streamon() 1099 mutex_unlock(&video->queue_lock); isp_video_streamon() 1135 mutex_lock(&video->queue_lock); isp_video_streamoff() 1137 mutex_unlock(&video->queue_lock); isp_video_streamoff() 1158 mutex_lock(&video->queue_lock); isp_video_streamoff() 1160 mutex_unlock(&video->queue_lock); isp_video_streamoff() 1293 mutex_lock(&video->queue_lock); isp_video_release() 1295 mutex_unlock(&video->queue_lock); isp_video_release() 1315 mutex_lock(&video->queue_lock); isp_video_poll() 1317 mutex_unlock(&video->queue_lock); isp_video_poll() 1382 mutex_init(&video->queue_lock); omap3isp_video_init() 1406 mutex_destroy(&video->queue_lock); omap3isp_video_cleanup()
|
H A D | ispvideo.h | 176 struct mutex queue_lock; /* protects the queue */ member in struct:isp_video
|
/linux-4.1.27/drivers/mmc/card/ |
H A D | queue.c | 62 spin_lock_irq(q->queue_lock); mmc_queue_thread() 66 spin_unlock_irq(q->queue_lock); mmc_queue_thread() 337 spin_lock_irqsave(q->queue_lock, flags); mmc_cleanup_queue() 340 spin_unlock_irqrestore(q->queue_lock, flags); mmc_cleanup_queue() 423 spin_lock_irqsave(q->queue_lock, flags); mmc_queue_suspend() 425 spin_unlock_irqrestore(q->queue_lock, flags); mmc_queue_suspend() 445 spin_lock_irqsave(q->queue_lock, flags); mmc_queue_resume() 447 spin_unlock_irqrestore(q->queue_lock, flags); mmc_queue_resume()
|
H A D | block.c | 1591 spin_lock_irq(q->queue_lock); mmc_blk_prep_packed_list() 1593 spin_unlock_irq(q->queue_lock); mmc_blk_prep_packed_list() 1628 spin_lock_irq(q->queue_lock); mmc_blk_prep_packed_list() 1630 spin_unlock_irq(q->queue_lock); mmc_blk_prep_packed_list() 1819 spin_lock_irq(q->queue_lock); mmc_blk_revert_packed_req() 1821 spin_unlock_irq(q->queue_lock); mmc_blk_revert_packed_req()
|
/linux-4.1.27/drivers/block/rsxx/ |
H A D | dma.c | 282 spin_lock_bh(&ctrl->queue_lock); rsxx_requeue_dma() 285 spin_unlock_bh(&ctrl->queue_lock); rsxx_requeue_dma() 383 spin_lock(&ctrl->queue_lock); dma_engine_stalled() 385 spin_unlock(&ctrl->queue_lock); dma_engine_stalled() 411 spin_lock_bh(&ctrl->queue_lock); rsxx_issue_dmas() 413 spin_unlock_bh(&ctrl->queue_lock); rsxx_issue_dmas() 416 spin_unlock_bh(&ctrl->queue_lock); rsxx_issue_dmas() 422 spin_lock_bh(&ctrl->queue_lock); rsxx_issue_dmas() 426 spin_unlock_bh(&ctrl->queue_lock); rsxx_issue_dmas() 586 spin_lock_bh(&ctrl->queue_lock); rsxx_dma_done() 589 spin_unlock_bh(&ctrl->queue_lock); rsxx_dma_done() 756 spin_lock_bh(&card->ctrl[i].queue_lock); 759 spin_unlock_bh(&card->ctrl[i].queue_lock); 838 spin_lock_init(&ctrl->queue_lock); rsxx_dma_ctrl_init() 1020 spin_lock_bh(&ctrl->queue_lock); rsxx_dma_destroy() 1022 spin_unlock_bh(&ctrl->queue_lock); rsxx_dma_destroy() 1076 spin_lock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_save_issued_dmas() 1082 spin_unlock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_save_issued_dmas()
|
H A D | core.c | 606 spin_lock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_failure() 610 spin_unlock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_failure() 713 spin_lock(&card->ctrl[i].queue_lock); rsxx_slot_reset() 715 spin_unlock(&card->ctrl[i].queue_lock); rsxx_slot_reset() 718 spin_unlock(&card->ctrl[i].queue_lock); rsxx_slot_reset()
|
H A D | rsxx_priv.h | 109 spinlock_t queue_lock; member in struct:rsxx_dma_ctrl
|
/linux-4.1.27/net/sunrpc/ |
H A D | cache.c | 725 static DEFINE_SPINLOCK(queue_lock); 770 spin_lock(&queue_lock); cache_read() 779 spin_unlock(&queue_lock); cache_read() 788 spin_unlock(&queue_lock); cache_read() 799 spin_lock(&queue_lock); cache_read() 801 spin_unlock(&queue_lock); cache_read() 811 spin_lock(&queue_lock); cache_read() 813 spin_unlock(&queue_lock); cache_read() 820 spin_lock(&queue_lock); cache_read() 825 spin_unlock(&queue_lock); cache_read() 830 spin_unlock(&queue_lock); cache_read() 929 spin_lock(&queue_lock); cache_poll() 937 spin_unlock(&queue_lock); cache_poll() 952 spin_lock(&queue_lock); cache_ioctl() 965 spin_unlock(&queue_lock); cache_ioctl() 987 spin_lock(&queue_lock); cache_open() 989 spin_unlock(&queue_lock); cache_open() 1001 spin_lock(&queue_lock); cache_release() 1014 spin_unlock(&queue_lock); cache_release() 1035 spin_lock(&queue_lock); cache_dequeue() 1048 spin_unlock(&queue_lock); cache_dequeue() 1182 spin_lock(&queue_lock); sunrpc_cache_pipe_upcall() 1188 spin_unlock(&queue_lock); sunrpc_cache_pipe_upcall()
|
/linux-4.1.27/drivers/ide/ |
H A D | ide-pm.c | 196 spin_lock_irqsave(q->queue_lock, flags); ide_complete_pm_rq() 201 spin_unlock_irqrestore(q->queue_lock, flags); ide_complete_pm_rq() 244 spin_lock_irqsave(q->queue_lock, flags); ide_check_pm_state() 246 spin_unlock_irqrestore(q->queue_lock, flags); ide_check_pm_state()
|
H A D | ide-io.c | 448 spin_lock_irqsave(q->queue_lock, flags); ide_requeue_and_plug() 450 spin_unlock_irqrestore(q->queue_lock, flags); ide_requeue_and_plug() 465 spin_unlock_irq(q->queue_lock); do_ide_request() 511 spin_lock_irq(q->queue_lock); do_ide_request() 519 spin_unlock_irq(q->queue_lock); do_ide_request() 565 spin_lock_irq(q->queue_lock); do_ide_request() 572 spin_lock_irq(q->queue_lock); do_ide_request()
|
/linux-4.1.27/drivers/scsi/esas2r/ |
H A D | esas2r_int.c | 218 spin_lock_irqsave(&a->queue_lock, flags); esas2r_get_outbound_responses() 228 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_get_outbound_responses() 235 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_get_outbound_responses() 295 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_get_outbound_responses() 344 spin_lock_irqsave(&a->queue_lock, flags); esas2r_do_deferred_processes() 376 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_do_deferred_processes() 399 spin_lock_irqsave(&a->queue_lock, flags); esas2r_process_adapter_reset() 440 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_process_adapter_reset() 458 spin_lock_irqsave(&a->queue_lock, flags); esas2r_process_bus_reset() 467 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_process_bus_reset()
|
H A D | esas2r_disc.c | 475 spin_lock_irqsave(&a->queue_lock, flags); esas2r_disc_start_request() 483 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_disc_start_request() 1167 spin_lock_irqsave(&a->queue_lock, flags); esas2r_disc_fix_curr_requests() 1183 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_disc_fix_curr_requests()
|
H A D | esas2r_io.c | 92 spin_lock_irqsave(&a->queue_lock, flags); esas2r_start_request() 102 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_start_request() 779 spin_lock_irqsave(&a->queue_lock, flags); esas2r_send_task_mgmt() 828 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_send_task_mgmt()
|
H A D | esas2r_init.c | 845 spin_lock_init(&a->queue_lock); esas2r_init_adapter_struct() 1331 spin_lock_irqsave(&a->queue_lock, flags); esas2r_init_msgs() 1333 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_init_msgs()
|
H A D | esas2r.h | 851 spinlock_t queue_lock; member in struct:esas2r_adapter 1400 spin_lock_irqsave(&a->queue_lock, flags); esas2r_start_ae_request() 1402 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_start_ae_request()
|
H A D | esas2r_main.c | 1060 spin_lock_irqsave(&a->queue_lock, flags); esas2r_eh_abort() 1074 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_eh_abort() 1081 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_eh_abort()
|
/linux-4.1.27/drivers/usb/host/ |
H A D | u132-hcd.c | 166 struct u132_spin queue_lock; member in struct:u132_endp 518 spin_lock_irqsave(&endp->queue_lock.slock, irqs); u132_hcd_giveback_urb() 523 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_giveback_urb() 532 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_giveback_urb() 557 spin_lock_irqsave(&endp->queue_lock.slock, irqs); u132_hcd_abandon_urb() 562 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_abandon_urb() 571 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_abandon_urb() 1875 spin_lock_init(&endp->queue_lock.slock); create_endpoint_and_queue_int() 1876 spin_lock_irqsave(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_int() 1879 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_int() 1930 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_int() 1974 spin_lock_init(&endp->queue_lock.slock); create_endpoint_and_queue_bulk() 1975 spin_lock_irqsave(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_bulk() 1978 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_bulk() 2029 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_bulk() 2070 spin_lock_init(&endp->queue_lock.slock); create_endpoint_and_queue_control() 2071 spin_lock_irqsave(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control() 2074 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control() 2118 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control() 2139 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control() 2276 spin_lock_irqsave(&endp->queue_lock.slock, u132_urb_enqueue() 2289 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_urb_enqueue() 2318 spin_lock_irqsave(&endp->queue_lock.slock, u132_urb_enqueue() 2331 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_urb_enqueue() 2367 spin_lock_irqsave(&endp->queue_lock.slock, u132_urb_enqueue() 2379 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_urb_enqueue() 2430 spin_lock_irqsave(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue() 2433 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue() 2442 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue() 2450 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue() 2453 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue() 2483 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_endp_urb_dequeue() 2491 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_endp_urb_dequeue() 2506 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue() 2514 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
|
/linux-4.1.27/drivers/media/usb/gspca/ |
H A D | gspca.c | 1173 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_s_fmt_vid_cap() 1199 mutex_unlock(&gspca_dev->queue_lock); vidioc_s_fmt_vid_cap() 1294 /* Needed for gspca_stream_off, always lock before queue_lock! */ dev_close() 1298 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) { dev_close() 1310 mutex_unlock(&gspca_dev->queue_lock); dev_close() 1385 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_reqbufs() 1433 mutex_unlock(&gspca_dev->queue_lock); vidioc_reqbufs() 1460 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_streamon() 1484 mutex_unlock(&gspca_dev->queue_lock); vidioc_streamon() 1497 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_streamoff() 1524 mutex_unlock(&gspca_dev->queue_lock); vidioc_streamoff() 1594 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) dev_mmap() 1646 mutex_unlock(&gspca_dev->queue_lock); dev_mmap() 1668 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) frame_ready() 1671 mutex_unlock(&gspca_dev->queue_lock); frame_ready() 1689 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_dqbuf() 1699 mutex_unlock(&gspca_dev->queue_lock); vidioc_dqbuf() 1713 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_dqbuf() 1737 mutex_unlock(&gspca_dev->queue_lock); vidioc_dqbuf() 1765 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_qbuf() 1804 mutex_unlock(&gspca_dev->queue_lock); vidioc_qbuf() 1879 if (mutex_lock_interruptible(&gspca_dev->queue_lock) != 0) { dev_poll() 1885 mutex_unlock(&gspca_dev->queue_lock); dev_poll() 2073 mutex_init(&gspca_dev->queue_lock); gspca_dev_probe2()
|
H A D | gspca.h | 151 /* protected by queue_lock */ 197 struct mutex queue_lock; /* ISOC queue protection */ member in struct:gspca_dev 211 /* (*) These variables are proteced by both usb_lock and queue_lock,
|
/linux-4.1.27/drivers/media/platform/ |
H A D | timblogiw.c | 69 spinlock_t queue_lock; /* mutual exclusion */ member in struct:timblogiw_fh 124 spin_lock(&fh->queue_lock); timblogiw_dma_cb() 146 spin_unlock(&fh->queue_lock); timblogiw_dma_cb() 564 spin_unlock_irq(&fh->queue_lock); buffer_queue() 570 spin_lock_irq(&fh->queue_lock); buffer_queue() 581 spin_lock_irq(&fh->queue_lock); buffer_queue() 662 spin_lock_init(&fh->queue_lock); timblogiw_open() 680 &timblogiw_video_qops, lw->dev, &fh->queue_lock, timblogiw_open()
|
/linux-4.1.27/include/linux/ |
H A D | connector.h | 40 spinlock_t queue_lock; member in struct:cn_queue_dev
|
H A D | srcu.h | 50 spinlock_t queue_lock; /* protect ->batch_queue, ->running */ member in struct:srcu_struct 91 .queue_lock = __SPIN_LOCK_UNLOCKED(name.queue_lock), \
|
H A D | blkdev.h | 390 * ->queue_lock. 393 spinlock_t *queue_lock; member in struct:request_queue 432 * queue_lock internally, e.g. scsi_request_fn(). 528 if (q->queue_lock) queue_lockdep_assert_held() 529 lockdep_assert_held(q->queue_lock); queue_lockdep_assert_held()
|
H A D | isdn.h | 252 spinlock_t queue_lock; /* lock to protect queue */ member in struct:isdn_net_dev_s
|
/linux-4.1.27/drivers/spi/ |
H A D | spi.c | 790 spin_lock_irqsave(&master->queue_lock, flags); __spi_pump_messages() 794 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 801 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 808 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 816 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 822 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 838 spin_lock_irqsave(&master->queue_lock, flags); __spi_pump_messages() 840 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 853 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages() 967 spin_lock_irqsave(&master->queue_lock, flags); spi_get_next_queued_message() 970 spin_unlock_irqrestore(&master->queue_lock, flags); spi_get_next_queued_message() 989 spin_lock_irqsave(&master->queue_lock, flags); spi_finalize_current_message() 991 spin_unlock_irqrestore(&master->queue_lock, flags); spi_finalize_current_message() 1003 spin_lock_irqsave(&master->queue_lock, flags); spi_finalize_current_message() 1007 spin_unlock_irqrestore(&master->queue_lock, flags); spi_finalize_current_message() 1021 spin_lock_irqsave(&master->queue_lock, flags); spi_start_queue() 1024 spin_unlock_irqrestore(&master->queue_lock, flags); spi_start_queue() 1030 spin_unlock_irqrestore(&master->queue_lock, flags); spi_start_queue() 1043 spin_lock_irqsave(&master->queue_lock, flags); spi_stop_queue() 1052 spin_unlock_irqrestore(&master->queue_lock, flags); spi_stop_queue() 1054 spin_lock_irqsave(&master->queue_lock, flags); spi_stop_queue() 1062 spin_unlock_irqrestore(&master->queue_lock, flags); spi_stop_queue() 1102 spin_lock_irqsave(&master->queue_lock, flags); __spi_queued_transfer() 1105 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_queued_transfer() 1115 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_queued_transfer() 1549 spin_lock_init(&master->queue_lock); spi_register_master()
|
H A D | spi-pl022.c | 352 * @queue_lock: spinlock to syncronise access to message queue
|
/linux-4.1.27/kernel/rcu/ |
H A D | srcu.c | 102 spin_lock_init(&sp->queue_lock); init_srcu_struct_fields() 395 spin_lock_irqsave(&sp->queue_lock, flags); call_srcu() 401 spin_unlock_irqrestore(&sp->queue_lock, flags); call_srcu() 428 spin_lock_irq(&sp->queue_lock); __synchronize_srcu() 433 spin_unlock_irq(&sp->queue_lock); __synchronize_srcu() 445 spin_unlock_irq(&sp->queue_lock); __synchronize_srcu() 548 spin_lock_irq(&sp->queue_lock); srcu_collect_new() 550 spin_unlock_irq(&sp->queue_lock); srcu_collect_new() 646 spin_lock_irq(&sp->queue_lock); srcu_reschedule() 654 spin_unlock_irq(&sp->queue_lock); srcu_reschedule()
|
/linux-4.1.27/drivers/s390/block/ |
H A D | dasd_ioctl.c | 161 spin_lock(&block->queue_lock); dasd_ioctl_abortio() 167 spin_unlock(&block->queue_lock); dasd_ioctl_abortio() 169 spin_lock(&block->queue_lock); dasd_ioctl_abortio() 172 spin_unlock(&block->queue_lock); dasd_ioctl_abortio()
|
H A D | dasd.c | 161 spin_lock_init(&block->queue_lock); dasd_alloc_block() 1800 spin_lock_bh(&block->queue_lock); list_for_each_safe() 1822 spin_unlock_bh(&block->queue_lock); list_for_each_safe() 2758 spin_lock(&block->queue_lock); dasd_block_tasklet() 2761 spin_unlock(&block->queue_lock); dasd_block_tasklet() 2769 spin_lock(&block->queue_lock); dasd_block_tasklet() 2774 spin_unlock(&block->queue_lock); dasd_block_tasklet() 2798 spin_lock_irqsave(&block->queue_lock, flags); _dasd_requeue_request() 2801 spin_unlock_irqrestore(&block->queue_lock, flags); _dasd_requeue_request() 2818 spin_lock_bh(&block->queue_lock); dasd_flush_block_queue() 2839 spin_unlock_bh(&block->queue_lock); dasd_flush_block_queue() 2846 spin_lock_bh(&block->queue_lock); dasd_flush_block_queue() 2848 spin_unlock_bh(&block->queue_lock); dasd_flush_block_queue() 2891 spin_lock(&block->queue_lock); do_dasd_request() 2896 spin_unlock(&block->queue_lock); do_dasd_request() 2926 spin_lock(&block->queue_lock); dasd_times_out() 2970 spin_unlock(&block->queue_lock); dasd_times_out()
|
H A D | dasd_int.h | 499 spinlock_t queue_lock; member in struct:dasd_block
|
/linux-4.1.27/include/linux/mtd/ |
H A D | blktrans.h | 50 spinlock_t queue_lock; member in struct:mtd_blktrans_dev
|
/linux-4.1.27/drivers/net/wireless/mwifiex/ |
H A D | init.c | 329 spin_lock_irqsave(&adapter->queue_lock, dev_queue_flags); mwifiex_wake_up_net_dev_queue() 338 spin_unlock_irqrestore(&adapter->queue_lock, dev_queue_flags); mwifiex_wake_up_net_dev_queue() 350 spin_lock_irqsave(&adapter->queue_lock, dev_queue_flags); mwifiex_stop_net_dev_queue() 359 spin_unlock_irqrestore(&adapter->queue_lock, dev_queue_flags); mwifiex_stop_net_dev_queue() 456 spin_lock_init(&adapter->queue_lock); mwifiex_init_lock_list()
|
H A D | main.h | 882 spinlock_t queue_lock; /* lock for tx queues */ member in struct:mwifiex_adapter
|
/linux-4.1.27/drivers/scsi/ |
H A D | scsi_lib.c | 170 spin_lock_irqsave(q->queue_lock, flags); __scsi_queue_insert() 173 spin_unlock_irqrestore(q->queue_lock, flags); __scsi_queue_insert() 429 * drops the queue_lock and can add us back to the scsi_starved_list_run() 534 spin_lock_irqsave(q->queue_lock, flags); scsi_requeue_command() 539 spin_unlock_irqrestore(q->queue_lock, flags); scsi_requeue_command() 728 spin_lock_irqsave(q->queue_lock, flags); scsi_end_request() 730 spin_unlock_irqrestore(q->queue_lock, flags); scsi_end_request() 1405 * Called with the queue_lock held. 1767 __releases(q->queue_lock) 1768 __acquires(q->queue_lock) 1807 spin_unlock_irq(q->queue_lock); 1859 spin_lock_irq(q->queue_lock); 1862 spin_lock_irq(q->queue_lock); 1873 * must return with queue_lock held. 1879 spin_lock_irq(q->queue_lock); 2965 spin_lock_irqsave(q->queue_lock, flags); scsi_internal_device_block() 2967 spin_unlock_irqrestore(q->queue_lock, flags); scsi_internal_device_block() 3017 spin_lock_irqsave(q->queue_lock, flags); scsi_internal_device_unblock() 3019 spin_unlock_irqrestore(q->queue_lock, flags); scsi_internal_device_unblock()
|
H A D | scsi_transport_srp.c | 414 spin_lock_irq(q->queue_lock); shost_for_each_device() 416 spin_unlock_irq(q->queue_lock); shost_for_each_device()
|
H A D | scsi_transport_fc.c | 3954 spin_unlock_irq(q->queue_lock); fc_bsg_request_handler() 3956 spin_lock_irq(q->queue_lock); fc_bsg_request_handler() 3960 spin_unlock_irq(q->queue_lock); fc_bsg_request_handler() 3966 spin_lock_irq(q->queue_lock); fc_bsg_request_handler() 3979 spin_lock_irq(q->queue_lock); fc_bsg_request_handler() 3983 /* the dispatch routines will unlock the queue_lock */ fc_bsg_request_handler() 3995 spin_lock_irq(q->queue_lock); fc_bsg_request_handler() 3998 spin_unlock_irq(q->queue_lock); fc_bsg_request_handler() 4000 spin_lock_irq(q->queue_lock); fc_bsg_request_handler()
|
H A D | scsi.c | 670 spin_lock_irqsave(sdev->request_queue->queue_lock, flags); scsi_change_queue_depth() 689 spin_unlock_irqrestore(sdev->request_queue->queue_lock, flags); scsi_change_queue_depth()
|
H A D | scsi_transport_sas.c | 178 spin_unlock_irq(q->queue_lock); sas_smp_request() 186 spin_lock_irq(q->queue_lock); sas_smp_request()
|
/linux-4.1.27/drivers/media/usb/usbvision/ |
H A D | usbvision-video.c | 817 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); vidioc_qbuf() 819 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); vidioc_qbuf() 841 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); vidioc_dqbuf() 845 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); vidioc_dqbuf() 1017 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); usbvision_read() 1019 spin_unlock_irqrestore(&usbvision->queue_lock, usbvision_read() 1036 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); usbvision_read() 1040 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); usbvision_read()
|
H A D | usbvision.h | 409 spinlock_t queue_lock; /* spinlock for protecting mods on inqueue and outqueue */ member in struct:usb_usbvision
|
H A D | usbvision-core.c | 1175 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); usbvision_parse_data() 1178 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); usbvision_parse_data() 1792 spin_lock_init(&usbvision->queue_lock); usbvision_frames_alloc()
|
/linux-4.1.27/drivers/media/usb/go7007/ |
H A D | go7007-priv.h | 233 struct mutex queue_lock; member in struct:go7007
|
H A D | go7007-usb.c | 1322 mutex_lock(&go->queue_lock); go7007_usb_disconnect() 1332 mutex_unlock(&go->queue_lock); go7007_usb_disconnect()
|
H A D | go7007-v4l2.c | 1106 mutex_init(&go->queue_lock); go7007_v4l2_init() 1116 go->vidq.lock = &go->queue_lock; go7007_v4l2_init()
|
/linux-4.1.27/drivers/staging/i2o/ |
H A D | i2o_block.c | 409 spin_lock_irqsave(q->queue_lock, flags); i2o_block_delayed_request_fn() 411 spin_unlock_irqrestore(q->queue_lock, flags); i2o_block_delayed_request_fn() 436 spin_lock_irqsave(q->queue_lock, flags); i2o_block_end_request() 445 spin_unlock_irqrestore(q->queue_lock, flags); i2o_block_end_request()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_83xx_hw.c | 3824 spin_lock(&mbx->queue_lock); qlcnic_83xx_flush_mbx_queue() 3835 spin_unlock(&mbx->queue_lock); qlcnic_83xx_flush_mbx_queue() 3871 spin_lock(&mbx->queue_lock); qlcnic_83xx_dequeue_mbx_cmd() 3876 spin_unlock(&mbx->queue_lock); qlcnic_83xx_dequeue_mbx_cmd() 3941 spin_lock(&mbx->queue_lock); qlcnic_83xx_enqueue_mbx_cmd() 3949 spin_unlock(&mbx->queue_lock); qlcnic_83xx_enqueue_mbx_cmd() 4043 spin_lock(&mbx->queue_lock); qlcnic_83xx_mailbox_worker() 4046 spin_unlock(&mbx->queue_lock); qlcnic_83xx_mailbox_worker() 4051 spin_unlock(&mbx->queue_lock); qlcnic_83xx_mailbox_worker() 4097 spin_lock_init(&mbx->queue_lock); qlcnic_83xx_init_mailbox_work()
|
H A D | qlcnic.h | 1100 spinlock_t queue_lock; /* Mailbox queue lock */ member in struct:qlcnic_mailbox
|
/linux-4.1.27/drivers/scsi/libsas/ |
H A D | sas_ata.c | 597 spin_lock_irqsave(q->queue_lock, flags); sas_ata_task_abort() 599 spin_unlock_irqrestore(q->queue_lock, flags); sas_ata_task_abort()
|
H A D | sas_scsi_host.c | 947 spin_lock_irqsave(q->queue_lock, flags); sas_task_abort() 949 spin_unlock_irqrestore(q->queue_lock, flags); sas_task_abort()
|
/linux-4.1.27/arch/x86/platform/uv/ |
H A D | tlb_uv.c | 717 spin_lock(&hmaster->queue_lock); destination_plugged() 719 spin_unlock(&hmaster->queue_lock); destination_plugged() 739 spin_lock(&hmaster->queue_lock); destination_timeout() 741 spin_unlock(&hmaster->queue_lock); destination_timeout() 1904 spin_lock_init(&bcp->queue_lock); for_each_present_cpu()
|
/linux-4.1.27/drivers/md/ |
H A D | dm.c | 1157 spin_lock_irqsave(q->queue_lock, flags); old_requeue_request() 1160 spin_unlock_irqrestore(q->queue_lock, flags); old_requeue_request() 1194 spin_lock_irqsave(q->queue_lock, flags); old_stop_queue() 1196 spin_unlock_irqrestore(q->queue_lock, flags); old_stop_queue() 1211 spin_lock_irqsave(q->queue_lock, flags); old_start_queue() 1214 spin_unlock_irqrestore(q->queue_lock, flags); old_start_queue() 2096 * ->pending within a single queue_lock not to increment the dm_request_fn()
|
H A D | dm-table.c | 1700 spin_lock_irqsave(queue->queue_lock, flags); dm_table_run_md_queue_async() 1702 spin_unlock_irqrestore(queue->queue_lock, flags); dm_table_run_md_queue_async()
|
/linux-4.1.27/kernel/ |
H A D | futex.c | 163 * Refer to the comment in queue_lock(). 1752 static inline struct futex_hash_bucket *queue_lock(struct futex_q *q) 1762 * waiting for the spinlock. This is safe as all queue_lock() 2145 *hb = queue_lock(q); futex_wait_setup() 2296 hb = queue_lock(&q); futex_lock_pi()
|
/linux-4.1.27/arch/x86/include/asm/uv/ |
H A D | uv_bau.h | 619 spinlock_t queue_lock; member in struct:bau_control
|
/linux-4.1.27/include/linux/spi/ |
H A D | spi.h | 253 * @queue_lock: spinlock to syncronise access to message queue 420 spinlock_t queue_lock; member in struct:spi_master
|
/linux-4.1.27/drivers/ata/ |
H A D | libata-eh.c | 985 spin_lock_irqsave(q->queue_lock, flags); ata_qc_schedule_eh() 987 spin_unlock_irqrestore(q->queue_lock, flags); ata_qc_schedule_eh()
|
/linux-4.1.27/drivers/block/drbd/ |
H A D | drbd_main.c | 2778 q->queue_lock = &resource->req_lock; drbd_create_device()
|