Searched refs:queue_lock (Results 1 - 90 of 90) sorted by relevance

/linux-4.1.27/drivers/connector/
H A Dcn_queue.c85 spin_lock_bh(&dev->queue_lock); cn_queue_add_callback()
94 spin_unlock_bh(&dev->queue_lock); cn_queue_add_callback()
112 spin_lock_bh(&dev->queue_lock); cn_queue_del_callback()
120 spin_unlock_bh(&dev->queue_lock); cn_queue_del_callback()
137 spin_lock_init(&dev->queue_lock); cn_queue_alloc_dev()
148 spin_lock_bh(&dev->queue_lock); cn_queue_free_dev()
151 spin_unlock_bh(&dev->queue_lock); cn_queue_free_dev()
H A Dconnector.c88 spin_lock_bh(&dev->cbdev->queue_lock); cn_netlink_send_mult()
97 spin_unlock_bh(&dev->cbdev->queue_lock); cn_netlink_send_mult()
156 spin_lock_bh(&dev->cbdev->queue_lock); cn_call_callback()
164 spin_unlock_bh(&dev->cbdev->queue_lock); cn_call_callback()
248 spin_lock_bh(&dev->queue_lock); cn_proc_show()
257 spin_unlock_bh(&dev->queue_lock); cn_proc_show()
/linux-4.1.27/drivers/scsi/arm/
H A Dqueue.c63 spin_lock_init(&queue->queue_lock); queue_initialise()
113 spin_lock_irqsave(&queue->queue_lock, flags); __queue_add()
133 spin_unlock_irqrestore(&queue->queue_lock, flags); __queue_add()
167 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_exclude()
176 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_exclude()
192 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove()
195 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove()
216 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_tgtluntag()
225 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_tgtluntag()
242 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_all_target()
248 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_all_target()
266 spin_lock_irqsave(&queue->queue_lock, flags); queue_probetgtlun()
274 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_probetgtlun()
292 spin_lock_irqsave(&queue->queue_lock, flags); queue_remove_cmd()
301 spin_unlock_irqrestore(&queue->queue_lock, flags); queue_remove_cmd()
H A Dqueue.h16 spinlock_t queue_lock; member in struct:__anon8658
/linux-4.1.27/drivers/isdn/i4l/
H A Disdn_net.h84 spin_lock_irqsave(&nd->queue_lock, flags); isdn_net_get_locked_lp()
95 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_get_locked_lp()
100 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_get_locked_lp()
112 spin_lock_irqsave(&nd->queue_lock, flags); isdn_net_add_to_bundle()
123 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_add_to_bundle()
138 spin_lock_irqsave(&master_lp->netdev->queue_lock, flags); isdn_net_rm_from_bundle()
150 spin_unlock_irqrestore(&master_lp->netdev->queue_lock, flags); isdn_net_rm_from_bundle()
H A Disdn_net.c128 spin_lock_irqsave(&nd->queue_lock, flags); isdn_net_device_busy()
132 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_device_busy()
137 spin_unlock_irqrestore(&nd->queue_lock, flags); isdn_net_device_busy()
2593 spin_lock_init(&netdev->queue_lock); isdn_net_new()
/linux-4.1.27/block/
H A Dblk-ioc.c60 lockdep_assert_held(q->queue_lock); ioc_destroy_icq()
68 * under queue_lock. If it's not pointing to @icq now, it never ioc_destroy_icq()
107 if (spin_trylock(q->queue_lock)) { ioc_release_fn()
109 spin_unlock(q->queue_lock); ioc_release_fn()
141 * already be holding a queue_lock. Do it asynchronously from wq. put_io_context()
185 if (spin_trylock(icq->q->queue_lock)) { put_io_context_active()
187 spin_unlock(icq->q->queue_lock); put_io_context_active()
221 lockdep_assert_held(q->queue_lock); ioc_clear_queue()
315 * with @q->queue_lock held.
321 lockdep_assert_held(q->queue_lock); ioc_lookup_icq()
380 spin_lock_irq(q->queue_lock); ioc_create_icq()
396 spin_unlock_irq(q->queue_lock); ioc_create_icq()
H A Dblk-cgroup.c125 * holding @q->queue_lock and lookup hint is updated on success.
138 * hint can only be updated under queue_lock as otherwise @blkg __blkg_lookup()
140 * responsible for grabbing queue_lock if @update_hint. __blkg_lookup()
145 lockdep_assert_held(q->queue_lock); __blkg_lookup()
185 lockdep_assert_held(q->queue_lock); blkg_create()
265 * should be called under RCU read lock and @q->queue_lock.
277 lockdep_assert_held(q->queue_lock); blkg_lookup_create()
315 lockdep_assert_held(blkg->q->queue_lock); blkg_destroy()
336 * under queue_lock. If it's not pointing to @blkg now, it never blkg_destroy()
368 lockdep_assert_held(q->queue_lock); blkg_destroy_all()
524 spin_lock_irq(blkg->q->queue_lock); blkcg_print_blkgs()
527 spin_unlock_irq(blkg->q->queue_lock); blkcg_print_blkgs()
637 lockdep_assert_held(pd->blkg->q->queue_lock); blkg_stat_recursive_sum()
671 lockdep_assert_held(pd->blkg->q->queue_lock); blkg_rwstat_recursive_sum()
707 __acquires(rcu) __acquires(disk->queue->queue_lock) __acquires()
727 spin_lock_irq(disk->queue->queue_lock); __acquires()
737 spin_unlock_irq(disk->queue->queue_lock); __acquires()
767 __releases(ctx->disk->queue->queue_lock) __releases(rcu) __releases()
769 spin_unlock_irq(ctx->disk->queue->queue_lock); __releases()
805 if (spin_trylock(q->queue_lock)) { blkcg_css_offline()
807 spin_unlock(q->queue_lock); blkcg_css_offline()
875 lockdep_assert_held(q->queue_lock); blkcg_drain_queue()
895 spin_lock_irq(q->queue_lock); blkcg_exit_queue()
897 spin_unlock_irq(q->queue_lock); blkcg_exit_queue()
987 spin_lock_irq(q->queue_lock); blkcg_activate_policy()
1008 spin_unlock_irq(q->queue_lock); blkcg_activate_policy()
1024 spin_lock_irq(q->queue_lock); blkcg_activate_policy()
1049 spin_unlock_irq(q->queue_lock); blkcg_activate_policy()
1075 spin_lock_irq(q->queue_lock); blkcg_deactivate_policy()
1098 spin_unlock_irq(q->queue_lock); blkcg_deactivate_policy()
H A Dblk-exec.c79 spin_lock_irq(q->queue_lock); blk_execute_rq_nowait()
85 spin_unlock_irq(q->queue_lock); blk_execute_rq_nowait()
94 spin_unlock_irq(q->queue_lock); blk_execute_rq_nowait()
H A Dblk-core.c163 spin_lock_irq(q->queue_lock); blk_delay_work()
165 spin_unlock_irq(q->queue_lock); blk_delay_work()
333 spin_lock_irqsave(q->queue_lock, flags); blk_run_queue()
335 spin_unlock_irqrestore(q->queue_lock, flags); blk_run_queue()
355 __releases(q->queue_lock)
356 __acquires(q->queue_lock)
360 lockdep_assert_held(q->queue_lock);
406 spin_unlock_irq(q->queue_lock);
410 spin_lock_irq(q->queue_lock);
439 spin_lock_irq(q->queue_lock); blk_queue_bypass_start()
442 spin_unlock_irq(q->queue_lock); blk_queue_bypass_start()
450 spin_lock_irq(q->queue_lock); blk_queue_bypass_start()
452 spin_unlock_irq(q->queue_lock); blk_queue_bypass_start()
468 spin_lock_irq(q->queue_lock); blk_queue_bypass_end()
472 spin_unlock_irq(q->queue_lock); blk_queue_bypass_end()
504 spinlock_t *lock = q->queue_lock; blk_cleanup_queue()
551 if (q->queue_lock != &q->__queue_lock) blk_cleanup_queue()
552 q->queue_lock = &q->__queue_lock; blk_cleanup_queue()
650 * By default initialize queue_lock to internal lock and driver can blk_alloc_queue_node()
653 q->queue_lock = &q->__queue_lock; blk_alloc_queue_node()
760 q->queue_lock = lock; blk_init_allocated_queue()
866 * congestion status, wake up any waiters. Called under q->queue_lock.
888 spin_lock_irq(q->queue_lock); blk_update_nr_requests()
921 spin_unlock_irq(q->queue_lock);
970 * Must be called with @q->queue_lock held and,
971 * Returns ERR_PTR on failure, with @q->queue_lock held.
972 * Returns request pointer on success, with @q->queue_lock *not held*.
1042 * Also, lookup icq while holding queue_lock. If it doesn't exist, __get_request()
1043 * it will be created after releasing queue_lock. __get_request()
1054 spin_unlock_irq(q->queue_lock); __get_request()
1108 spin_lock_irq(q->queue_lock); __get_request()
1110 spin_unlock_irq(q->queue_lock); __get_request()
1121 spin_lock_irq(q->queue_lock); __get_request()
1147 * Must be called with @q->queue_lock held and,
1148 * Returns ERR_PTR on failure, with @q->queue_lock held.
1149 * Returns request pointer on success, with @q->queue_lock *not held*.
1176 spin_unlock_irq(q->queue_lock); get_request()
1186 spin_lock_irq(q->queue_lock); get_request()
1202 spin_lock_irq(q->queue_lock); blk_old_get_request()
1205 spin_unlock_irq(q->queue_lock); blk_old_get_request()
1206 /* q->queue_lock is unlocked at this point */ blk_old_get_request()
1424 spin_lock_irqsave(q->queue_lock, flags); blk_put_request()
1426 spin_unlock_irqrestore(q->queue_lock, flags); blk_put_request()
1519 * going through @q->queue_lock. As such it's more of an issuing mechanism
1604 spin_lock_irq(q->queue_lock); blk_queue_bio()
1617 spin_lock_irq(q->queue_lock); blk_queue_bio()
1684 spin_lock_irq(q->queue_lock); blk_queue_bio()
1688 spin_unlock_irq(q->queue_lock); blk_queue_bio()
2074 spin_lock_irqsave(q->queue_lock, flags); blk_insert_cloned_request()
2076 spin_unlock_irqrestore(q->queue_lock, flags); blk_insert_cloned_request()
2092 spin_unlock_irqrestore(q->queue_lock, flags); blk_insert_cloned_request()
2112 * queue_lock must be held.
2254 * queue_lock must be held.
2377 * queue_lock must be held.
2409 * queue_lock must be held.
2654 spin_lock_irqsave(q->queue_lock, flags); blk_end_bidi_request()
2656 spin_unlock_irqrestore(q->queue_lock, flags); blk_end_bidi_request()
3072 __releases(q->queue_lock)
3080 spin_unlock(q->queue_lock);
3165 spin_lock(q->queue_lock); blk_flush_plug_list()
3261 spin_lock_irq(q->queue_lock); blk_pre_runtime_suspend()
3268 spin_unlock_irq(q->queue_lock); blk_pre_runtime_suspend()
3288 spin_lock_irq(q->queue_lock); blk_post_runtime_suspend()
3295 spin_unlock_irq(q->queue_lock); blk_post_runtime_suspend()
3312 spin_lock_irq(q->queue_lock); blk_pre_runtime_resume()
3314 spin_unlock_irq(q->queue_lock); blk_pre_runtime_resume()
3334 spin_lock_irq(q->queue_lock); blk_post_runtime_resume()
3343 spin_unlock_irq(q->queue_lock); blk_post_runtime_resume()
H A Dblk-timeout.c60 spin_lock_irq(q->queue_lock); part_timeout_store()
65 spin_unlock_irq(q->queue_lock); part_timeout_store()
137 spin_lock_irqsave(q->queue_lock, flags); blk_rq_timed_out_timer()
145 spin_unlock_irqrestore(q->queue_lock, flags); blk_rq_timed_out_timer()
H A Dnoop-iosched.c80 spin_lock_irq(q->queue_lock); noop_init_queue()
82 spin_unlock_irq(q->queue_lock); noop_init_queue()
H A Dblk-sysfs.c179 spin_lock_irq(q->queue_lock); queue_max_sectors_store()
181 spin_unlock_irq(q->queue_lock); queue_max_sectors_store()
212 spin_lock_irq(q->queue_lock); \
217 spin_unlock_irq(q->queue_lock); \
241 spin_lock_irq(q->queue_lock); queue_nomerges_store()
248 spin_unlock_irq(q->queue_lock); queue_nomerges_store()
272 spin_lock_irq(q->queue_lock); queue_rq_affinity_store()
283 spin_unlock_irq(q->queue_lock); queue_rq_affinity_store()
507 spin_lock_irq(q->queue_lock); blk_release_queue()
509 spin_unlock_irq(q->queue_lock); blk_release_queue()
H A Dbsg-lib.c177 spin_unlock_irq(q->queue_lock); bsg_request_fn()
183 spin_lock_irq(q->queue_lock); bsg_request_fn()
189 spin_lock_irq(q->queue_lock); bsg_request_fn()
194 spin_unlock_irq(q->queue_lock); bsg_request_fn()
196 spin_lock_irq(q->queue_lock); bsg_request_fn()
H A Delevator.c582 lockdep_assert_held(q->queue_lock); elv_drain_elevator()
678 spin_lock_irqsave(q->queue_lock, flags); elv_add_request()
680 spin_unlock_irqrestore(q->queue_lock, flags); elv_add_request()
911 spin_lock_irq(q->queue_lock); elevator_switch()
913 spin_unlock_irq(q->queue_lock); elevator_switch()
H A Dblk-throttle.c1206 spin_lock_irq(q->queue_lock); throtl_pending_timer_fn()
1226 spin_unlock_irq(q->queue_lock); throtl_pending_timer_fn()
1228 spin_lock_irq(q->queue_lock); throtl_pending_timer_fn()
1250 spin_unlock_irq(q->queue_lock); throtl_pending_timer_fn()
1274 spin_lock_irq(q->queue_lock); blk_throtl_dispatch_work_fn()
1278 spin_unlock_irq(q->queue_lock); blk_throtl_dispatch_work_fn()
1394 * We're already holding queue_lock and know @tg is valid. Let's tg_set_conf()
1514 spin_lock_irq(q->queue_lock); blk_throtl_bio()
1582 spin_unlock_irq(q->queue_lock); blk_throtl_bio()
1625 __releases(q->queue_lock) __acquires(q->queue_lock)
1649 spin_unlock_irq(q->queue_lock);
1657 spin_lock_irq(q->queue_lock);
H A Dblk-flush.c160 * spin_lock_irq(q->queue_lock or fq->mq_flush_lock)
279 * spin_lock_irq(q->queue_lock or fq->mq_flush_lock)
369 * spin_lock_irq(q->queue_lock) in !mq case
H A Ddeadline-iosched.c366 spin_lock_irq(q->queue_lock); deadline_init_queue()
368 spin_unlock_irq(q->queue_lock); deadline_init_queue()
H A Dcfq-iosched.c487 /* This should be called with the queue_lock held. */ cfqg_stats_update_group_wait_time()
502 /* This should be called with the queue_lock held. */ cfqg_stats_set_start_group_wait_time()
516 /* This should be called with the queue_lock held. */ cfqg_stats_end_empty_time()
743 lockdep_assert_held(cfqg_to_blkg(cfqg)->q->queue_lock); cfqg_stats_xfer_dead()
2845 * Must be called with the queue_lock held.
3611 spin_unlock_irq(cfqd->queue->queue_lock); cfq_find_alloc_queue()
3615 spin_lock_irq(cfqd->queue->queue_lock); cfq_find_alloc_queue()
4226 spin_lock_irq(q->queue_lock); cfq_set_request()
4262 spin_unlock_irq(q->queue_lock); cfq_set_request()
4272 spin_lock_irq(q->queue_lock); cfq_kick_queue()
4274 spin_unlock_irq(q->queue_lock); cfq_kick_queue()
4289 spin_lock_irqsave(cfqd->queue->queue_lock, flags); cfq_idle_slice_timer()
4330 spin_unlock_irqrestore(cfqd->queue->queue_lock, flags); cfq_idle_slice_timer()
4361 spin_lock_irq(q->queue_lock); cfq_exit_queue()
4368 spin_unlock_irq(q->queue_lock); cfq_exit_queue()
4399 spin_lock_irq(q->queue_lock); cfq_init_queue()
4401 spin_unlock_irq(q->queue_lock); cfq_init_queue()
4443 spin_lock_irq(q->queue_lock); cfq_init_queue()
4446 spin_unlock_irq(q->queue_lock); cfq_init_queue()
H A Dblk-cgroup.h321 * under queue_lock. This function is guaranteed to return non-%NULL
360 * queue_lock.
398 * Should be used under queue_lock.
H A Dblk-mq.c117 spin_lock_irq(q->queue_lock); blk_mq_freeze_queue_start()
119 spin_unlock_irq(q->queue_lock); blk_mq_freeze_queue_start()
148 spin_lock_irq(q->queue_lock); blk_mq_unfreeze_queue()
151 spin_unlock_irq(q->queue_lock); blk_mq_unfreeze_queue()
/linux-4.1.27/drivers/scsi/device_handler/
H A Dscsi_dh.c126 spin_lock_irq(sdev->request_queue->queue_lock); scsi_dh_handler_attach()
128 spin_unlock_irq(sdev->request_queue->queue_lock); scsi_dh_handler_attach()
141 spin_lock_irq(sdev->request_queue->queue_lock); __detach_handler()
143 spin_unlock_irq(sdev->request_queue->queue_lock); __detach_handler()
418 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_activate()
421 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_activate()
437 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_activate()
470 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_set_params()
476 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_set_params()
514 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_attach()
518 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_attach()
543 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_detach()
547 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_detach()
575 spin_lock_irqsave(q->queue_lock, flags); scsi_dh_attached_handler_name()
579 spin_unlock_irqrestore(q->queue_lock, flags); scsi_dh_attached_handler_name()
/linux-4.1.27/drivers/block/
H A Dnbd.c48 spinlock_t queue_lock; member in struct:nbd_device
108 spin_lock_irqsave(q->queue_lock, flags); nbd_end_request()
110 spin_unlock_irqrestore(q->queue_lock, flags); nbd_end_request()
290 spin_lock(&nbd->queue_lock); nbd_find_request()
295 spin_unlock(&nbd->queue_lock); nbd_find_request()
298 spin_unlock(&nbd->queue_lock); nbd_find_request()
491 spin_lock(&nbd->queue_lock); nbd_handle_req()
493 spin_unlock(&nbd->queue_lock); nbd_handle_req()
523 spin_lock_irq(&nbd->queue_lock); nbd_thread()
527 spin_unlock_irq(&nbd->queue_lock); nbd_thread()
543 __releases(q->queue_lock) __acquires(q->queue_lock)
550 spin_unlock_irq(q->queue_lock);
564 spin_lock_irq(q->queue_lock);
568 spin_lock_irq(&nbd->queue_lock);
570 spin_unlock_irq(&nbd->queue_lock);
574 spin_lock_irq(q->queue_lock);
844 spin_lock_init(&nbd_dev[i].queue_lock); nbd_init()
H A DDAC960.c528 spin_unlock_irq(&Controller->queue_lock); DAC960_WaitForCommand()
530 spin_lock_irq(&Controller->queue_lock); DAC960_WaitForCommand()
790 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_ExecuteCommand()
792 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_ExecuteCommand()
2043 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration()
2045 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration()
2088 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration()
2090 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ReadDeviceConfiguration()
2536 RequestQueue = blk_init_queue(DAC960_RequestFunction,&Controller->queue_lock); DAC960_RegisterBlockDevice()
2780 spin_lock_init(&Controller->queue_lock); DAC960_DetectController()
3119 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_FinalizeController()
3121 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_FinalizeController()
5271 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_GEM_InterruptHandler()
5295 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_GEM_InterruptHandler()
5312 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_BA_InterruptHandler()
5336 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_BA_InterruptHandler()
5354 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_LP_InterruptHandler()
5378 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_LP_InterruptHandler()
5396 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_LA_InterruptHandler()
5416 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_LA_InterruptHandler()
5434 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_PG_InterruptHandler()
5454 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_PG_InterruptHandler()
5471 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_PD_InterruptHandler()
5488 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_PD_InterruptHandler()
5509 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_P_InterruptHandler()
5561 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_P_InterruptHandler()
5628 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction()
5636 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction()
5686 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction()
5694 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_MonitoringTimerFunction()
5971 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand()
5974 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand()
6165 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand()
6167 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V1_ExecuteUserCommand()
6234 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand()
6237 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand()
6428 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand()
6430 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_V2_ExecuteUserCommand()
6765 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6771 spin_unlock_irq(&Controller->queue_lock); DAC960_gam_ioctl()
6775 spin_lock_irq(&Controller->queue_lock); DAC960_gam_ioctl()
6779 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6790 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6793 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6804 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6806 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6911 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6914 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6962 spin_lock_irqsave(&Controller->queue_lock, flags); DAC960_gam_ioctl()
6964 spin_unlock_irqrestore(&Controller->queue_lock, flags); DAC960_gam_ioctl()
H A Dhd.c540 spin_lock_irq(hd_queue->queue_lock); hd_times_out()
551 spin_unlock_irq(hd_queue->queue_lock); hd_times_out()
680 spin_lock(hd_queue->queue_lock); hd_interrupt()
688 spin_unlock(hd_queue->queue_lock); hd_interrupt()
H A Dnull_blk.c348 spin_unlock_irq(q->queue_lock); null_request_fn()
350 spin_lock_irq(q->queue_lock); null_request_fn()
H A Dnvme-core.c604 spin_lock_irqsave(req->q->queue_lock, flags); req_completion()
607 spin_unlock_irqrestore(req->q->queue_lock, flags); req_completion()
2593 spin_lock(ns->queue->queue_lock); nvme_freeze_queues()
2595 spin_unlock(ns->queue->queue_lock); nvme_freeze_queues()
H A Dfloppy.c2251 spin_lock_irqsave(q->queue_lock, flags); request_done()
2253 spin_unlock_irqrestore(q->queue_lock, flags); request_done()
2265 spin_lock_irqsave(q->queue_lock, flags); request_done()
2267 spin_unlock_irqrestore(q->queue_lock, flags); request_done()
H A Dumem.c898 card->queue->queue_lock = &card->lock; mm_pci_probe()
H A Dpktcdvd.c2220 spin_lock_irq(q->queue_lock); pkt_open_dev()
2222 spin_unlock_irq(q->queue_lock); pkt_open_dev()
H A Dcciss.c3314 spin_unlock_irq(q->queue_lock); do_cciss_request()
3428 spin_lock_irq(q->queue_lock); do_cciss_request()
H A DDAC960.h2325 spinlock_t queue_lock; member in struct:DAC960_Controller
/linux-4.1.27/drivers/mtd/
H A Dmtd_blkdevs.c137 spin_lock_irq(rq->queue_lock); mtd_blktrans_work()
145 spin_unlock_irq(rq->queue_lock); mtd_blktrans_work()
149 spin_lock_irq(rq->queue_lock); mtd_blktrans_work()
160 spin_unlock_irq(rq->queue_lock); mtd_blktrans_work()
166 spin_lock_irq(rq->queue_lock); mtd_blktrans_work()
174 spin_unlock_irq(rq->queue_lock); mtd_blktrans_work()
409 spin_lock_init(&new->queue_lock); add_mtd_blktrans_dev()
410 new->rq = blk_init_queue(mtd_blktrans_request, &new->queue_lock); add_mtd_blktrans_dev()
481 spin_lock_irqsave(&old->queue_lock, flags); del_mtd_blktrans_dev()
484 spin_unlock_irqrestore(&old->queue_lock, flags); del_mtd_blktrans_dev()
/linux-4.1.27/drivers/media/platform/omap3isp/
H A Dispvideo.c572 mutex_lock(&video->queue_lock); omap3isp_video_resume()
574 mutex_unlock(&video->queue_lock); omap3isp_video_resume()
827 mutex_lock(&video->queue_lock); isp_video_reqbufs()
829 mutex_unlock(&video->queue_lock); isp_video_reqbufs()
841 mutex_lock(&video->queue_lock); isp_video_querybuf()
843 mutex_unlock(&video->queue_lock); isp_video_querybuf()
855 mutex_lock(&video->queue_lock); isp_video_qbuf()
857 mutex_unlock(&video->queue_lock); isp_video_qbuf()
869 mutex_lock(&video->queue_lock); isp_video_dqbuf()
871 mutex_unlock(&video->queue_lock); isp_video_dqbuf()
1072 mutex_lock(&video->queue_lock); isp_video_streamon()
1074 mutex_unlock(&video->queue_lock); isp_video_streamon()
1097 mutex_lock(&video->queue_lock); isp_video_streamon()
1099 mutex_unlock(&video->queue_lock); isp_video_streamon()
1135 mutex_lock(&video->queue_lock); isp_video_streamoff()
1137 mutex_unlock(&video->queue_lock); isp_video_streamoff()
1158 mutex_lock(&video->queue_lock); isp_video_streamoff()
1160 mutex_unlock(&video->queue_lock); isp_video_streamoff()
1293 mutex_lock(&video->queue_lock); isp_video_release()
1295 mutex_unlock(&video->queue_lock); isp_video_release()
1315 mutex_lock(&video->queue_lock); isp_video_poll()
1317 mutex_unlock(&video->queue_lock); isp_video_poll()
1382 mutex_init(&video->queue_lock); omap3isp_video_init()
1406 mutex_destroy(&video->queue_lock); omap3isp_video_cleanup()
H A Dispvideo.h176 struct mutex queue_lock; /* protects the queue */ member in struct:isp_video
/linux-4.1.27/drivers/mmc/card/
H A Dqueue.c62 spin_lock_irq(q->queue_lock); mmc_queue_thread()
66 spin_unlock_irq(q->queue_lock); mmc_queue_thread()
337 spin_lock_irqsave(q->queue_lock, flags); mmc_cleanup_queue()
340 spin_unlock_irqrestore(q->queue_lock, flags); mmc_cleanup_queue()
423 spin_lock_irqsave(q->queue_lock, flags); mmc_queue_suspend()
425 spin_unlock_irqrestore(q->queue_lock, flags); mmc_queue_suspend()
445 spin_lock_irqsave(q->queue_lock, flags); mmc_queue_resume()
447 spin_unlock_irqrestore(q->queue_lock, flags); mmc_queue_resume()
H A Dblock.c1591 spin_lock_irq(q->queue_lock); mmc_blk_prep_packed_list()
1593 spin_unlock_irq(q->queue_lock); mmc_blk_prep_packed_list()
1628 spin_lock_irq(q->queue_lock); mmc_blk_prep_packed_list()
1630 spin_unlock_irq(q->queue_lock); mmc_blk_prep_packed_list()
1819 spin_lock_irq(q->queue_lock); mmc_blk_revert_packed_req()
1821 spin_unlock_irq(q->queue_lock); mmc_blk_revert_packed_req()
/linux-4.1.27/drivers/block/rsxx/
H A Ddma.c282 spin_lock_bh(&ctrl->queue_lock); rsxx_requeue_dma()
285 spin_unlock_bh(&ctrl->queue_lock); rsxx_requeue_dma()
383 spin_lock(&ctrl->queue_lock); dma_engine_stalled()
385 spin_unlock(&ctrl->queue_lock); dma_engine_stalled()
411 spin_lock_bh(&ctrl->queue_lock); rsxx_issue_dmas()
413 spin_unlock_bh(&ctrl->queue_lock); rsxx_issue_dmas()
416 spin_unlock_bh(&ctrl->queue_lock); rsxx_issue_dmas()
422 spin_lock_bh(&ctrl->queue_lock); rsxx_issue_dmas()
426 spin_unlock_bh(&ctrl->queue_lock); rsxx_issue_dmas()
586 spin_lock_bh(&ctrl->queue_lock); rsxx_dma_done()
589 spin_unlock_bh(&ctrl->queue_lock); rsxx_dma_done()
756 spin_lock_bh(&card->ctrl[i].queue_lock);
759 spin_unlock_bh(&card->ctrl[i].queue_lock);
838 spin_lock_init(&ctrl->queue_lock); rsxx_dma_ctrl_init()
1020 spin_lock_bh(&ctrl->queue_lock); rsxx_dma_destroy()
1022 spin_unlock_bh(&ctrl->queue_lock); rsxx_dma_destroy()
1076 spin_lock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_save_issued_dmas()
1082 spin_unlock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_save_issued_dmas()
H A Dcore.c606 spin_lock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_failure()
610 spin_unlock_bh(&card->ctrl[i].queue_lock); rsxx_eeh_failure()
713 spin_lock(&card->ctrl[i].queue_lock); rsxx_slot_reset()
715 spin_unlock(&card->ctrl[i].queue_lock); rsxx_slot_reset()
718 spin_unlock(&card->ctrl[i].queue_lock); rsxx_slot_reset()
H A Drsxx_priv.h109 spinlock_t queue_lock; member in struct:rsxx_dma_ctrl
/linux-4.1.27/net/sunrpc/
H A Dcache.c725 static DEFINE_SPINLOCK(queue_lock);
770 spin_lock(&queue_lock); cache_read()
779 spin_unlock(&queue_lock); cache_read()
788 spin_unlock(&queue_lock); cache_read()
799 spin_lock(&queue_lock); cache_read()
801 spin_unlock(&queue_lock); cache_read()
811 spin_lock(&queue_lock); cache_read()
813 spin_unlock(&queue_lock); cache_read()
820 spin_lock(&queue_lock); cache_read()
825 spin_unlock(&queue_lock); cache_read()
830 spin_unlock(&queue_lock); cache_read()
929 spin_lock(&queue_lock); cache_poll()
937 spin_unlock(&queue_lock); cache_poll()
952 spin_lock(&queue_lock); cache_ioctl()
965 spin_unlock(&queue_lock); cache_ioctl()
987 spin_lock(&queue_lock); cache_open()
989 spin_unlock(&queue_lock); cache_open()
1001 spin_lock(&queue_lock); cache_release()
1014 spin_unlock(&queue_lock); cache_release()
1035 spin_lock(&queue_lock); cache_dequeue()
1048 spin_unlock(&queue_lock); cache_dequeue()
1182 spin_lock(&queue_lock); sunrpc_cache_pipe_upcall()
1188 spin_unlock(&queue_lock); sunrpc_cache_pipe_upcall()
/linux-4.1.27/drivers/ide/
H A Dide-pm.c196 spin_lock_irqsave(q->queue_lock, flags); ide_complete_pm_rq()
201 spin_unlock_irqrestore(q->queue_lock, flags); ide_complete_pm_rq()
244 spin_lock_irqsave(q->queue_lock, flags); ide_check_pm_state()
246 spin_unlock_irqrestore(q->queue_lock, flags); ide_check_pm_state()
H A Dide-io.c448 spin_lock_irqsave(q->queue_lock, flags); ide_requeue_and_plug()
450 spin_unlock_irqrestore(q->queue_lock, flags); ide_requeue_and_plug()
465 spin_unlock_irq(q->queue_lock); do_ide_request()
511 spin_lock_irq(q->queue_lock); do_ide_request()
519 spin_unlock_irq(q->queue_lock); do_ide_request()
565 spin_lock_irq(q->queue_lock); do_ide_request()
572 spin_lock_irq(q->queue_lock); do_ide_request()
/linux-4.1.27/drivers/scsi/esas2r/
H A Desas2r_int.c218 spin_lock_irqsave(&a->queue_lock, flags); esas2r_get_outbound_responses()
228 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_get_outbound_responses()
235 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_get_outbound_responses()
295 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_get_outbound_responses()
344 spin_lock_irqsave(&a->queue_lock, flags); esas2r_do_deferred_processes()
376 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_do_deferred_processes()
399 spin_lock_irqsave(&a->queue_lock, flags); esas2r_process_adapter_reset()
440 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_process_adapter_reset()
458 spin_lock_irqsave(&a->queue_lock, flags); esas2r_process_bus_reset()
467 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_process_bus_reset()
H A Desas2r_disc.c475 spin_lock_irqsave(&a->queue_lock, flags); esas2r_disc_start_request()
483 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_disc_start_request()
1167 spin_lock_irqsave(&a->queue_lock, flags); esas2r_disc_fix_curr_requests()
1183 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_disc_fix_curr_requests()
H A Desas2r_io.c92 spin_lock_irqsave(&a->queue_lock, flags); esas2r_start_request()
102 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_start_request()
779 spin_lock_irqsave(&a->queue_lock, flags); esas2r_send_task_mgmt()
828 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_send_task_mgmt()
H A Desas2r_init.c845 spin_lock_init(&a->queue_lock); esas2r_init_adapter_struct()
1331 spin_lock_irqsave(&a->queue_lock, flags); esas2r_init_msgs()
1333 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_init_msgs()
H A Desas2r.h851 spinlock_t queue_lock; member in struct:esas2r_adapter
1400 spin_lock_irqsave(&a->queue_lock, flags); esas2r_start_ae_request()
1402 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_start_ae_request()
H A Desas2r_main.c1060 spin_lock_irqsave(&a->queue_lock, flags); esas2r_eh_abort()
1074 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_eh_abort()
1081 spin_unlock_irqrestore(&a->queue_lock, flags); esas2r_eh_abort()
/linux-4.1.27/drivers/usb/host/
H A Du132-hcd.c166 struct u132_spin queue_lock; member in struct:u132_endp
518 spin_lock_irqsave(&endp->queue_lock.slock, irqs); u132_hcd_giveback_urb()
523 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_giveback_urb()
532 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_giveback_urb()
557 spin_lock_irqsave(&endp->queue_lock.slock, irqs); u132_hcd_abandon_urb()
562 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_abandon_urb()
571 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_hcd_abandon_urb()
1875 spin_lock_init(&endp->queue_lock.slock); create_endpoint_and_queue_int()
1876 spin_lock_irqsave(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_int()
1879 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_int()
1930 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_int()
1974 spin_lock_init(&endp->queue_lock.slock); create_endpoint_and_queue_bulk()
1975 spin_lock_irqsave(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_bulk()
1978 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_bulk()
2029 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_bulk()
2070 spin_lock_init(&endp->queue_lock.slock); create_endpoint_and_queue_control()
2071 spin_lock_irqsave(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control()
2074 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control()
2118 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control()
2139 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); create_endpoint_and_queue_control()
2276 spin_lock_irqsave(&endp->queue_lock.slock, u132_urb_enqueue()
2289 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_urb_enqueue()
2318 spin_lock_irqsave(&endp->queue_lock.slock, u132_urb_enqueue()
2331 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_urb_enqueue()
2367 spin_lock_irqsave(&endp->queue_lock.slock, u132_urb_enqueue()
2379 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_urb_enqueue()
2430 spin_lock_irqsave(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
2433 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
2442 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
2450 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
2453 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
2483 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_endp_urb_dequeue()
2491 spin_unlock_irqrestore(&endp->queue_lock.slock, u132_endp_urb_dequeue()
2506 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
2514 spin_unlock_irqrestore(&endp->queue_lock.slock, irqs); u132_endp_urb_dequeue()
/linux-4.1.27/drivers/media/usb/gspca/
H A Dgspca.c1173 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_s_fmt_vid_cap()
1199 mutex_unlock(&gspca_dev->queue_lock); vidioc_s_fmt_vid_cap()
1294 /* Needed for gspca_stream_off, always lock before queue_lock! */ dev_close()
1298 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) { dev_close()
1310 mutex_unlock(&gspca_dev->queue_lock); dev_close()
1385 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_reqbufs()
1433 mutex_unlock(&gspca_dev->queue_lock); vidioc_reqbufs()
1460 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_streamon()
1484 mutex_unlock(&gspca_dev->queue_lock); vidioc_streamon()
1497 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_streamoff()
1524 mutex_unlock(&gspca_dev->queue_lock); vidioc_streamoff()
1594 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) dev_mmap()
1646 mutex_unlock(&gspca_dev->queue_lock); dev_mmap()
1668 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) frame_ready()
1671 mutex_unlock(&gspca_dev->queue_lock); frame_ready()
1689 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_dqbuf()
1699 mutex_unlock(&gspca_dev->queue_lock); vidioc_dqbuf()
1713 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_dqbuf()
1737 mutex_unlock(&gspca_dev->queue_lock); vidioc_dqbuf()
1765 if (mutex_lock_interruptible(&gspca_dev->queue_lock)) vidioc_qbuf()
1804 mutex_unlock(&gspca_dev->queue_lock); vidioc_qbuf()
1879 if (mutex_lock_interruptible(&gspca_dev->queue_lock) != 0) { dev_poll()
1885 mutex_unlock(&gspca_dev->queue_lock); dev_poll()
2073 mutex_init(&gspca_dev->queue_lock); gspca_dev_probe2()
H A Dgspca.h151 /* protected by queue_lock */
197 struct mutex queue_lock; /* ISOC queue protection */ member in struct:gspca_dev
211 /* (*) These variables are proteced by both usb_lock and queue_lock,
/linux-4.1.27/drivers/media/platform/
H A Dtimblogiw.c69 spinlock_t queue_lock; /* mutual exclusion */ member in struct:timblogiw_fh
124 spin_lock(&fh->queue_lock); timblogiw_dma_cb()
146 spin_unlock(&fh->queue_lock); timblogiw_dma_cb()
564 spin_unlock_irq(&fh->queue_lock); buffer_queue()
570 spin_lock_irq(&fh->queue_lock); buffer_queue()
581 spin_lock_irq(&fh->queue_lock); buffer_queue()
662 spin_lock_init(&fh->queue_lock); timblogiw_open()
680 &timblogiw_video_qops, lw->dev, &fh->queue_lock, timblogiw_open()
/linux-4.1.27/include/linux/
H A Dconnector.h40 spinlock_t queue_lock; member in struct:cn_queue_dev
H A Dsrcu.h50 spinlock_t queue_lock; /* protect ->batch_queue, ->running */ member in struct:srcu_struct
91 .queue_lock = __SPIN_LOCK_UNLOCKED(name.queue_lock), \
H A Dblkdev.h390 * ->queue_lock.
393 spinlock_t *queue_lock; member in struct:request_queue
432 * queue_lock internally, e.g. scsi_request_fn().
528 if (q->queue_lock) queue_lockdep_assert_held()
529 lockdep_assert_held(q->queue_lock); queue_lockdep_assert_held()
H A Disdn.h252 spinlock_t queue_lock; /* lock to protect queue */ member in struct:isdn_net_dev_s
/linux-4.1.27/drivers/spi/
H A Dspi.c790 spin_lock_irqsave(&master->queue_lock, flags); __spi_pump_messages()
794 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
801 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
808 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
816 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
822 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
838 spin_lock_irqsave(&master->queue_lock, flags); __spi_pump_messages()
840 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
853 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_pump_messages()
967 spin_lock_irqsave(&master->queue_lock, flags); spi_get_next_queued_message()
970 spin_unlock_irqrestore(&master->queue_lock, flags); spi_get_next_queued_message()
989 spin_lock_irqsave(&master->queue_lock, flags); spi_finalize_current_message()
991 spin_unlock_irqrestore(&master->queue_lock, flags); spi_finalize_current_message()
1003 spin_lock_irqsave(&master->queue_lock, flags); spi_finalize_current_message()
1007 spin_unlock_irqrestore(&master->queue_lock, flags); spi_finalize_current_message()
1021 spin_lock_irqsave(&master->queue_lock, flags); spi_start_queue()
1024 spin_unlock_irqrestore(&master->queue_lock, flags); spi_start_queue()
1030 spin_unlock_irqrestore(&master->queue_lock, flags); spi_start_queue()
1043 spin_lock_irqsave(&master->queue_lock, flags); spi_stop_queue()
1052 spin_unlock_irqrestore(&master->queue_lock, flags); spi_stop_queue()
1054 spin_lock_irqsave(&master->queue_lock, flags); spi_stop_queue()
1062 spin_unlock_irqrestore(&master->queue_lock, flags); spi_stop_queue()
1102 spin_lock_irqsave(&master->queue_lock, flags); __spi_queued_transfer()
1105 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_queued_transfer()
1115 spin_unlock_irqrestore(&master->queue_lock, flags); __spi_queued_transfer()
1549 spin_lock_init(&master->queue_lock); spi_register_master()
H A Dspi-pl022.c352 * @queue_lock: spinlock to syncronise access to message queue
/linux-4.1.27/kernel/rcu/
H A Dsrcu.c102 spin_lock_init(&sp->queue_lock); init_srcu_struct_fields()
395 spin_lock_irqsave(&sp->queue_lock, flags); call_srcu()
401 spin_unlock_irqrestore(&sp->queue_lock, flags); call_srcu()
428 spin_lock_irq(&sp->queue_lock); __synchronize_srcu()
433 spin_unlock_irq(&sp->queue_lock); __synchronize_srcu()
445 spin_unlock_irq(&sp->queue_lock); __synchronize_srcu()
548 spin_lock_irq(&sp->queue_lock); srcu_collect_new()
550 spin_unlock_irq(&sp->queue_lock); srcu_collect_new()
646 spin_lock_irq(&sp->queue_lock); srcu_reschedule()
654 spin_unlock_irq(&sp->queue_lock); srcu_reschedule()
/linux-4.1.27/drivers/s390/block/
H A Ddasd_ioctl.c161 spin_lock(&block->queue_lock); dasd_ioctl_abortio()
167 spin_unlock(&block->queue_lock); dasd_ioctl_abortio()
169 spin_lock(&block->queue_lock); dasd_ioctl_abortio()
172 spin_unlock(&block->queue_lock); dasd_ioctl_abortio()
H A Ddasd.c161 spin_lock_init(&block->queue_lock); dasd_alloc_block()
1800 spin_lock_bh(&block->queue_lock); list_for_each_safe()
1822 spin_unlock_bh(&block->queue_lock); list_for_each_safe()
2758 spin_lock(&block->queue_lock); dasd_block_tasklet()
2761 spin_unlock(&block->queue_lock); dasd_block_tasklet()
2769 spin_lock(&block->queue_lock); dasd_block_tasklet()
2774 spin_unlock(&block->queue_lock); dasd_block_tasklet()
2798 spin_lock_irqsave(&block->queue_lock, flags); _dasd_requeue_request()
2801 spin_unlock_irqrestore(&block->queue_lock, flags); _dasd_requeue_request()
2818 spin_lock_bh(&block->queue_lock); dasd_flush_block_queue()
2839 spin_unlock_bh(&block->queue_lock); dasd_flush_block_queue()
2846 spin_lock_bh(&block->queue_lock); dasd_flush_block_queue()
2848 spin_unlock_bh(&block->queue_lock); dasd_flush_block_queue()
2891 spin_lock(&block->queue_lock); do_dasd_request()
2896 spin_unlock(&block->queue_lock); do_dasd_request()
2926 spin_lock(&block->queue_lock); dasd_times_out()
2970 spin_unlock(&block->queue_lock); dasd_times_out()
H A Ddasd_int.h499 spinlock_t queue_lock; member in struct:dasd_block
/linux-4.1.27/include/linux/mtd/
H A Dblktrans.h50 spinlock_t queue_lock; member in struct:mtd_blktrans_dev
/linux-4.1.27/drivers/net/wireless/mwifiex/
H A Dinit.c329 spin_lock_irqsave(&adapter->queue_lock, dev_queue_flags); mwifiex_wake_up_net_dev_queue()
338 spin_unlock_irqrestore(&adapter->queue_lock, dev_queue_flags); mwifiex_wake_up_net_dev_queue()
350 spin_lock_irqsave(&adapter->queue_lock, dev_queue_flags); mwifiex_stop_net_dev_queue()
359 spin_unlock_irqrestore(&adapter->queue_lock, dev_queue_flags); mwifiex_stop_net_dev_queue()
456 spin_lock_init(&adapter->queue_lock); mwifiex_init_lock_list()
H A Dmain.h882 spinlock_t queue_lock; /* lock for tx queues */ member in struct:mwifiex_adapter
/linux-4.1.27/drivers/scsi/
H A Dscsi_lib.c170 spin_lock_irqsave(q->queue_lock, flags); __scsi_queue_insert()
173 spin_unlock_irqrestore(q->queue_lock, flags); __scsi_queue_insert()
429 * drops the queue_lock and can add us back to the scsi_starved_list_run()
534 spin_lock_irqsave(q->queue_lock, flags); scsi_requeue_command()
539 spin_unlock_irqrestore(q->queue_lock, flags); scsi_requeue_command()
728 spin_lock_irqsave(q->queue_lock, flags); scsi_end_request()
730 spin_unlock_irqrestore(q->queue_lock, flags); scsi_end_request()
1405 * Called with the queue_lock held.
1767 __releases(q->queue_lock)
1768 __acquires(q->queue_lock)
1807 spin_unlock_irq(q->queue_lock);
1859 spin_lock_irq(q->queue_lock);
1862 spin_lock_irq(q->queue_lock);
1873 * must return with queue_lock held.
1879 spin_lock_irq(q->queue_lock);
2965 spin_lock_irqsave(q->queue_lock, flags); scsi_internal_device_block()
2967 spin_unlock_irqrestore(q->queue_lock, flags); scsi_internal_device_block()
3017 spin_lock_irqsave(q->queue_lock, flags); scsi_internal_device_unblock()
3019 spin_unlock_irqrestore(q->queue_lock, flags); scsi_internal_device_unblock()
H A Dscsi_transport_srp.c414 spin_lock_irq(q->queue_lock); shost_for_each_device()
416 spin_unlock_irq(q->queue_lock); shost_for_each_device()
H A Dscsi_transport_fc.c3954 spin_unlock_irq(q->queue_lock); fc_bsg_request_handler()
3956 spin_lock_irq(q->queue_lock); fc_bsg_request_handler()
3960 spin_unlock_irq(q->queue_lock); fc_bsg_request_handler()
3966 spin_lock_irq(q->queue_lock); fc_bsg_request_handler()
3979 spin_lock_irq(q->queue_lock); fc_bsg_request_handler()
3983 /* the dispatch routines will unlock the queue_lock */ fc_bsg_request_handler()
3995 spin_lock_irq(q->queue_lock); fc_bsg_request_handler()
3998 spin_unlock_irq(q->queue_lock); fc_bsg_request_handler()
4000 spin_lock_irq(q->queue_lock); fc_bsg_request_handler()
H A Dscsi.c670 spin_lock_irqsave(sdev->request_queue->queue_lock, flags); scsi_change_queue_depth()
689 spin_unlock_irqrestore(sdev->request_queue->queue_lock, flags); scsi_change_queue_depth()
H A Dscsi_transport_sas.c178 spin_unlock_irq(q->queue_lock); sas_smp_request()
186 spin_lock_irq(q->queue_lock); sas_smp_request()
/linux-4.1.27/drivers/media/usb/usbvision/
H A Dusbvision-video.c817 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); vidioc_qbuf()
819 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); vidioc_qbuf()
841 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); vidioc_dqbuf()
845 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); vidioc_dqbuf()
1017 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); usbvision_read()
1019 spin_unlock_irqrestore(&usbvision->queue_lock, usbvision_read()
1036 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); usbvision_read()
1040 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); usbvision_read()
H A Dusbvision.h409 spinlock_t queue_lock; /* spinlock for protecting mods on inqueue and outqueue */ member in struct:usb_usbvision
H A Dusbvision-core.c1175 spin_lock_irqsave(&usbvision->queue_lock, lock_flags); usbvision_parse_data()
1178 spin_unlock_irqrestore(&usbvision->queue_lock, lock_flags); usbvision_parse_data()
1792 spin_lock_init(&usbvision->queue_lock); usbvision_frames_alloc()
/linux-4.1.27/drivers/media/usb/go7007/
H A Dgo7007-priv.h233 struct mutex queue_lock; member in struct:go7007
H A Dgo7007-usb.c1322 mutex_lock(&go->queue_lock); go7007_usb_disconnect()
1332 mutex_unlock(&go->queue_lock); go7007_usb_disconnect()
H A Dgo7007-v4l2.c1106 mutex_init(&go->queue_lock); go7007_v4l2_init()
1116 go->vidq.lock = &go->queue_lock; go7007_v4l2_init()
/linux-4.1.27/drivers/staging/i2o/
H A Di2o_block.c409 spin_lock_irqsave(q->queue_lock, flags); i2o_block_delayed_request_fn()
411 spin_unlock_irqrestore(q->queue_lock, flags); i2o_block_delayed_request_fn()
436 spin_lock_irqsave(q->queue_lock, flags); i2o_block_end_request()
445 spin_unlock_irqrestore(q->queue_lock, flags); i2o_block_end_request()
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/
H A Dqlcnic_83xx_hw.c3824 spin_lock(&mbx->queue_lock); qlcnic_83xx_flush_mbx_queue()
3835 spin_unlock(&mbx->queue_lock); qlcnic_83xx_flush_mbx_queue()
3871 spin_lock(&mbx->queue_lock); qlcnic_83xx_dequeue_mbx_cmd()
3876 spin_unlock(&mbx->queue_lock); qlcnic_83xx_dequeue_mbx_cmd()
3941 spin_lock(&mbx->queue_lock); qlcnic_83xx_enqueue_mbx_cmd()
3949 spin_unlock(&mbx->queue_lock); qlcnic_83xx_enqueue_mbx_cmd()
4043 spin_lock(&mbx->queue_lock); qlcnic_83xx_mailbox_worker()
4046 spin_unlock(&mbx->queue_lock); qlcnic_83xx_mailbox_worker()
4051 spin_unlock(&mbx->queue_lock); qlcnic_83xx_mailbox_worker()
4097 spin_lock_init(&mbx->queue_lock); qlcnic_83xx_init_mailbox_work()
H A Dqlcnic.h1100 spinlock_t queue_lock; /* Mailbox queue lock */ member in struct:qlcnic_mailbox
/linux-4.1.27/drivers/scsi/libsas/
H A Dsas_ata.c597 spin_lock_irqsave(q->queue_lock, flags); sas_ata_task_abort()
599 spin_unlock_irqrestore(q->queue_lock, flags); sas_ata_task_abort()
H A Dsas_scsi_host.c947 spin_lock_irqsave(q->queue_lock, flags); sas_task_abort()
949 spin_unlock_irqrestore(q->queue_lock, flags); sas_task_abort()
/linux-4.1.27/arch/x86/platform/uv/
H A Dtlb_uv.c717 spin_lock(&hmaster->queue_lock); destination_plugged()
719 spin_unlock(&hmaster->queue_lock); destination_plugged()
739 spin_lock(&hmaster->queue_lock); destination_timeout()
741 spin_unlock(&hmaster->queue_lock); destination_timeout()
1904 spin_lock_init(&bcp->queue_lock); for_each_present_cpu()
/linux-4.1.27/drivers/md/
H A Ddm.c1157 spin_lock_irqsave(q->queue_lock, flags); old_requeue_request()
1160 spin_unlock_irqrestore(q->queue_lock, flags); old_requeue_request()
1194 spin_lock_irqsave(q->queue_lock, flags); old_stop_queue()
1196 spin_unlock_irqrestore(q->queue_lock, flags); old_stop_queue()
1211 spin_lock_irqsave(q->queue_lock, flags); old_start_queue()
1214 spin_unlock_irqrestore(q->queue_lock, flags); old_start_queue()
2096 * ->pending within a single queue_lock not to increment the dm_request_fn()
H A Ddm-table.c1700 spin_lock_irqsave(queue->queue_lock, flags); dm_table_run_md_queue_async()
1702 spin_unlock_irqrestore(queue->queue_lock, flags); dm_table_run_md_queue_async()
/linux-4.1.27/kernel/
H A Dfutex.c163 * Refer to the comment in queue_lock().
1752 static inline struct futex_hash_bucket *queue_lock(struct futex_q *q)
1762 * waiting for the spinlock. This is safe as all queue_lock()
2145 *hb = queue_lock(q); futex_wait_setup()
2296 hb = queue_lock(&q); futex_lock_pi()
/linux-4.1.27/arch/x86/include/asm/uv/
H A Duv_bau.h619 spinlock_t queue_lock; member in struct:bau_control
/linux-4.1.27/include/linux/spi/
H A Dspi.h253 * @queue_lock: spinlock to syncronise access to message queue
420 spinlock_t queue_lock; member in struct:spi_master
/linux-4.1.27/drivers/ata/
H A Dlibata-eh.c985 spin_lock_irqsave(q->queue_lock, flags); ata_qc_schedule_eh()
987 spin_unlock_irqrestore(q->queue_lock, flags); ata_qc_schedule_eh()
/linux-4.1.27/drivers/block/drbd/
H A Ddrbd_main.c2778 q->queue_lock = &resource->req_lock; drbd_create_device()

Completed in 2392 milliseconds