Lines Matching refs:bdev
78 struct scm_blk_dev *bdev = scmrq->bdev; in scm_reserve_cluster() local
85 spin_lock(&bdev->lock); in scm_reserve_cluster()
86 list_for_each_entry(iter, &bdev->cluster_list, cluster.list) { in scm_reserve_cluster()
99 spin_unlock(&bdev->lock); in scm_reserve_cluster()
105 list_add(&scmrq->cluster.list, &bdev->cluster_list); in scm_reserve_cluster()
106 spin_unlock(&bdev->lock); in scm_reserve_cluster()
113 struct scm_blk_dev *bdev = scmrq->bdev; in scm_release_cluster() local
119 spin_lock_irqsave(&bdev->lock, flags); in scm_release_cluster()
121 spin_unlock_irqrestore(&bdev->lock, flags); in scm_release_cluster()
124 void scm_blk_dev_cluster_setup(struct scm_blk_dev *bdev) in scm_blk_dev_cluster_setup() argument
126 INIT_LIST_HEAD(&bdev->cluster_list); in scm_blk_dev_cluster_setup()
127 blk_queue_io_opt(bdev->rq, CLUSTER_SIZE); in scm_blk_dev_cluster_setup()
132 struct scm_blk_dev *bdev = scmrq->bdev; in scm_prepare_cluster_request() local
133 struct scm_device *scmdev = bdev->gendisk->private_data; in scm_prepare_cluster_request()
226 struct scm_blk_dev *bdev = scmrq->bdev; in scm_cluster_request_irq() local
239 spin_lock_irqsave(&bdev->rq_lock, flags); in scm_cluster_request_irq()
241 spin_unlock_irqrestore(&bdev->rq_lock, flags); in scm_cluster_request_irq()