Home
last modified time | relevance | path

Searched refs:blk_rq_pos (Results 1 – 62 of 62) sorted by relevance

/linux-4.1.27/drivers/s390/block/
Dscm_blk_cluster.c64 firstA = ((u64) blk_rq_pos(A) << 9) / CLUSTER_SIZE; in clusters_intersect()
65 lastA = (((u64) blk_rq_pos(A) << 9) + in clusters_intersect()
68 firstB = ((u64) blk_rq_pos(B) << 9) / CLUSTER_SIZE; in clusters_intersect()
69 lastB = (((u64) blk_rq_pos(B) << 9) + in clusters_intersect()
152 addr = scmdev->address + ((u64) blk_rq_pos(req) << 9); in scm_prepare_cluster_request()
177 addr < scmdev->address + ((u64) blk_rq_pos(req) << 9); in scm_prepare_cluster_request()
Ddasd_fba.c279 first_rec = blk_rq_pos(req) >> block->s2b_shift; in dasd_fba_build_cp()
281 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; in dasd_fba_build_cp()
316 block->bp_block, blk_rq_pos(req), blk_rq_sectors(req)); in dasd_fba_build_cp()
Ddasd_diag.c522 first_rec = blk_rq_pos(req) >> block->s2b_shift; in dasd_diag_build_cp()
524 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; in dasd_diag_build_cp()
Dscm_blk.c198 msb->scm_addr = scmdev->address + ((u64) blk_rq_pos(req) << 9); in scm_request_prepare()
Ddasd_eckd.c3251 first_rec = first_trk = blk_rq_pos(req) >> block->s2b_shift; in dasd_eckd_build_cp()
3254 (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift; in dasd_eckd_build_cp()
3330 start_padding_sectors = blk_rq_pos(req) % DASD_RAW_SECTORS_PER_TRACK; in dasd_raw_build_cp()
3331 end_sector_offset = (blk_rq_pos(req) + blk_rq_sectors(req)) % in dasd_raw_build_cp()
3345 first_trk = blk_rq_pos(req) / DASD_RAW_SECTORS_PER_TRACK; in dasd_raw_build_cp()
3346 last_trk = (blk_rq_pos(req) + blk_rq_sectors(req) - 1) / in dasd_raw_build_cp()
3471 recid = blk_rq_pos(req) >> cqr->block->s2b_shift; in dasd_eckd_free_cp()
/linux-4.1.27/block/
Delevator.c50 #define rq_hash_key(rq) (blk_rq_pos(rq) + blk_rq_sectors(rq))
307 if (blk_rq_pos(rq) < blk_rq_pos(__rq)) in elv_rb_add()
309 else if (blk_rq_pos(rq) >= blk_rq_pos(__rq)) in elv_rb_add()
334 if (sector < blk_rq_pos(rq)) in elv_rb_find()
336 else if (sector > blk_rq_pos(rq)) in elv_rb_find()
376 if (blk_rq_pos(rq) >= boundary) { in elv_dispatch_sort()
377 if (blk_rq_pos(pos) < boundary) in elv_dispatch_sort()
380 if (blk_rq_pos(pos) >= boundary) in elv_dispatch_sort()
383 if (blk_rq_pos(rq) >= blk_rq_pos(pos)) in elv_dispatch_sort()
485 __rq = elv_rqhash_find(q, blk_rq_pos(rq)); in elv_attempt_insert_merge()
Dblk-merge.c474 if (blk_rq_pos(req) + blk_rq_sectors(req) != blk_rq_pos(next)) in attempt_merge()
605 if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector) in blk_try_merge()
607 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector) in blk_try_merge()
Dcfq-iosched.c1096 s1 = blk_rq_pos(rq1); in cfq_choose_req()
1097 s2 = blk_rq_pos(rq2); in cfq_choose_req()
1228 return cfq_choose_req(cfqd, next, prev, blk_rq_pos(last)); in cfq_find_next_rq()
2120 if (sector > blk_rq_pos(cfqq->next_rq)) in cfq_prio_tree_lookup()
2122 else if (sector < blk_rq_pos(cfqq->next_rq)) in cfq_prio_tree_lookup()
2153 blk_rq_pos(cfqq->next_rq), &parent, &p); in cfq_prio_tree_add()
2307 cfqd->last_position = blk_rq_pos(rq) + blk_rq_sectors(rq); in cfq_activate_request()
2574 if (blk_rq_pos(rq) >= cfqd->last_position) in cfq_dist_from_last()
2575 return blk_rq_pos(rq) - cfqd->last_position; in cfq_dist_from_last()
2577 return cfqd->last_position - blk_rq_pos(rq); in cfq_dist_from_last()
[all …]
Dblk-core.c144 (unsigned long long)blk_rq_pos(rq), in blk_dump_rq_flags()
2219 part = disk_map_sector_rcu(rq->rq_disk, blk_rq_pos(rq)); in blk_account_io_start()
2495 (unsigned long long)blk_rq_pos(req)); in blk_update_request()
2935 dst->__sector = blk_rq_pos(src); in __blk_rq_prep_clone()
3061 (rqa->q == rqb->q && blk_rq_pos(rqa) < blk_rq_pos(rqb))); in plug_rq_cmp()
Ddeadline-iosched.c139 BUG_ON(sector != blk_rq_pos(__rq)); in deadline_merge()
Dblk-mq.c1100 blk_rq_pos(rqa) < blk_rq_pos(rqb))); in plug_ctx_cmp()
/linux-4.1.27/drivers/block/
Dmg_disk.c207 (unsigned int)blk_rq_pos(host->req)); in mg_dump_status()
494 if (mg_out(host, blk_rq_pos(req), blk_rq_sectors(req), in mg_read()
499 blk_rq_sectors(req), blk_rq_pos(req), bio_data(req->bio)); in mg_read()
530 if (mg_out(host, blk_rq_pos(req), rem, in mg_write()
537 rem, blk_rq_pos(req), bio_data(req->bio)); in mg_write()
588 blk_rq_pos(req), blk_rq_sectors(req) - 1, bio_data(req->bio)); in mg_read_intr()
627 blk_rq_pos(req), blk_rq_sectors(req), bio_data(req->bio)); in mg_write_intr()
743 sect_num = blk_rq_pos(req); in mg_request()
Dz2ram.c75 unsigned long start = blk_rq_pos(req) << 9; in do_z2_request()
82 (unsigned long long)blk_rq_pos(req), in do_z2_request()
Dswim3.c344 (long)blk_rq_pos(req), blk_rq_sectors(req), in start_request()
350 if (blk_rq_pos(req) >= fs->total_secs) { in start_request()
352 (long)blk_rq_pos(req), (long)fs->total_secs); in start_request()
378 fs->req_cyl = ((long)blk_rq_pos(req)) / fs->secpercyl; in start_request()
379 x = ((long)blk_rq_pos(req)) % fs->secpercyl; in start_request()
668 (long)blk_rq_pos(fs->cur_req)); in xfer_timeout()
788 (long)blk_rq_pos(req), err); in swim3_interrupt()
Dosdblk.c353 osd_req_write(or, &osdev->obj, blk_rq_pos(rq) * 512ULL, in osdblk_rq_fn()
356 osd_req_read(or, &osdev->obj, blk_rq_pos(rq) * 512ULL, in osdblk_rq_fn()
362 blk_rq_pos(rq) * 512ULL); in osdblk_rq_fn()
Dhd.c239 printk(", sector=%ld", blk_rq_pos(hd_req)); in dump_status()
470 req->rq_disk->disk_name, blk_rq_pos(req) + 1, in read_intr()
604 block = blk_rq_pos(req); in hd_request()
Dnbd.c239 request.from = cpu_to_be64((u64)blk_rq_pos(req) << 9); in nbd_send_req()
246 (unsigned long long)blk_rq_pos(req) << 9, blk_rq_bytes(req)); in nbd_send_req()
Dswim.c538 if (blk_rq_pos(req) >= fs->total_secs) in redo_fd_request()
550 err = floppy_read_sectors(fs, blk_rq_pos(req), in redo_fd_request()
Dataflop.c1428 fd_request ? blk_rq_pos(fd_request) : 0 )); in redo_fd_request()
1476 if (blk_rq_pos(fd_request) + 1 > UDT->blocks) { in redo_fd_request()
1486 ReqBlock = blk_rq_pos(fd_request); in redo_fd_request()
Dxen-blkfront.c444 ring_req->u.discard.sector_number = (blkif_sector_t)blk_rq_pos(req); in blkif_queue_request()
465 ring_req->u.indirect.sector_number = (blkif_sector_t)blk_rq_pos(req); in blkif_queue_request()
469 ring_req->u.rw.sector_number = (blkif_sector_t)blk_rq_pos(req); in blkif_queue_request()
635 req, req->cmd, (unsigned long)blk_rq_pos(req), in do_blkif_request()
Dxsysace.c659 (unsigned long long)blk_rq_pos(req), in ace_fsm_dostate()
666 ace_out32(ace, ACE_MPULBA, blk_rq_pos(req) & 0x0FFFFFFF); in ace_fsm_dostate()
Dfloppy.c2245 block = current_count_sectors + blk_rq_pos(req); in request_done()
2259 DRWE->first_error_sector = blk_rq_pos(req); in request_done()
2262 DRWE->last_error_sector = blk_rq_pos(req); in request_done()
2548 TRACK = (int)blk_rq_pos(current_req) / max_sector; in make_raw_rw_request()
2549 fsector_t = (int)blk_rq_pos(current_req) % max_sector; in make_raw_rw_request()
2904 current_req, (long)blk_rq_pos(current_req), current_req->cmd_type, in do_fd_request()
Dps3disk.c142 start_sector = blk_rq_pos(req) * priv->blocking_factor; in ps3disk_submit_request_sg()
Dsx8.c908 msg->lba = cpu_to_le32(blk_rq_pos(rq) & 0xffffffff); in carm_rq_fn()
909 tmp = (blk_rq_pos(rq) >> 16) >> 16; in carm_rq_fn()
Damiflop.c1398 blk_rq_pos(rq), cnt, in redo_fd_request()
1401 block = blk_rq_pos(rq) + cnt; in redo_fd_request()
Dnvme-core.c523 phys = nvme_block_nr(ns, blk_rq_pos(req)); in nvme_dif_remap()
729 range->slba = cpu_to_le64(nvme_block_nr(ns, blk_rq_pos(req))); in nvme_submit_discard()
783 cmnd->rw.slba = cpu_to_le64(nvme_block_nr(ns, blk_rq_pos(req))); in nvme_submit_iod()
797 nvme_block_nr(ns, blk_rq_pos(req))); in nvme_submit_iod()
Dcpqarray.c929 c->req.hdr.blk = blk_rq_pos(creq); in do_ida_request()
933 blk_rq_pos(creq), blk_rq_sectors(creq)); in do_ida_request()
Dvirtio_blk.c183 vbr->out_hdr.sector = cpu_to_virtio64(vblk->vdev, blk_rq_pos(vbr->req)); in virtio_queue_rq()
Dsunvdc.c492 desc->offset = (blk_rq_pos(req) << 9) / port->vdisk_block_size; in __send_request()
Dskd_main.c624 lba = (u32)blk_rq_pos(req); in skd_request_fn()
867 u32 lba = (u32)blk_rq_pos(req); in skd_end_request()
5326 u32 lba = (u32)blk_rq_pos(req); in skd_log_skreq()
Dloop.c397 pos = ((loff_t) blk_rq_pos(rq) << 9) + lo->lo_offset; in do_req_filebacked()
Dcciss.c3336 start_blk = blk_rq_pos(creq); in do_cciss_request()
3338 (int)blk_rq_pos(creq), (int)blk_rq_sectors(creq)); in do_cciss_request()
DDAC960.c3350 Command->BlockNumber = blk_rq_pos(Request); in DAC960_process_queue()
3443 Command->BlockNumber = blk_rq_pos(Request); in DAC960_queue_partial_rw()
Drbd.c3338 u64 offset = (u64)blk_rq_pos(rq) << SECTOR_SHIFT; in rbd_queue_workfn()
/linux-4.1.27/include/scsi/
Dscsi_cmnd.h294 return blk_rq_pos(scmd->request); in scsi_get_lba()
304 return blk_rq_pos(scmd->request) >> in scsi_prot_ref_tag()
/linux-4.1.27/include/trace/events/
Dblock.h82 0 : blk_rq_pos(rq);
161 __entry->sector = blk_rq_pos(rq);
195 0 : blk_rq_pos(rq);
654 __entry->sector = blk_rq_pos(rq);
/linux-4.1.27/drivers/scsi/
Du14-34f.c1309 flush_dev(SCpnt->device, blk_rq_pos(SCpnt->request), j, FALSE); in u14_34f_queuecommand_lck()
1608 if (blk_rq_pos(SCpnt->request) < minsec) in reorder()
1609 minsec = blk_rq_pos(SCpnt->request); in reorder()
1610 if (blk_rq_pos(SCpnt->request) > maxsec) in reorder()
1611 maxsec = blk_rq_pos(SCpnt->request); in reorder()
1613 sl[n] = blk_rq_pos(SCpnt->request); in reorder()
1670 blk_rq_pos(SCpnt->request), blk_rq_sectors(SCpnt->request), in reorder()
1799 flush_dev(SCpnt->device, blk_rq_pos(SCpnt->request), j, TRUE); in ihdlr()
Deata.c1816 flush_dev(SCpnt->device, blk_rq_pos(SCpnt->request), ha, 0); in eata2x_queuecommand_lck()
2133 if (blk_rq_pos(SCpnt->request) < minsec) in reorder()
2134 minsec = blk_rq_pos(SCpnt->request); in reorder()
2135 if (blk_rq_pos(SCpnt->request) > maxsec) in reorder()
2136 maxsec = blk_rq_pos(SCpnt->request); in reorder()
2138 sl[n] = blk_rq_pos(SCpnt->request); in reorder()
2229 n_ready, blk_rq_pos(SCpnt->request), in reorder()
2397 flush_dev(SCpnt->device, blk_rq_pos(SCpnt->request), ha, 1); in ihdlr()
Dsr.c358 blk_rq_pos(SCpnt->request)) << 9; in sr_done()
471 if (((unsigned int)blk_rq_pos(rq) % (s_size >> 9)) || in sr_init_command()
487 block = (unsigned int)blk_rq_pos(rq) / (s_size >> 9); in sr_init_command()
Dsd.c695 sector_t sector = blk_rq_pos(rq); in sd_setup_discard_cmnd()
821 sector_t sector = blk_rq_pos(rq); in sd_setup_write_same_cmnd()
889 sector_t block = blk_rq_pos(rq); in sd_setup_read_write_cmnd()
1602 u64 start_lba = blk_rq_pos(scmd->request); in sd_completed_bytes()
1603 u64 end_lba = blk_rq_pos(scmd->request) + (scsi_bufflen(scmd) / 512); in sd_completed_bytes()
/linux-4.1.27/include/linux/
Delevator.h201 #define rq_end_sector(rq) (blk_rq_pos(rq) + blk_rq_sectors(rq))
Dblkdev.h876 static inline sector_t blk_rq_pos(const struct request *rq) in blk_rq_pos() function
939 return min(blk_max_size_offset(q, blk_rq_pos(rq)), in blk_rq_get_max_sectors()
/linux-4.1.27/drivers/ide/
Dide-lib.c100 (unsigned long long)blk_rq_pos(rq)); in ide_dump_ata_error()
Dide-floppy.c259 if (((long)blk_rq_pos(rq) % floppy->bs_factor) || in ide_floppy_do_request()
Dide-io.c369 return drv->do_request(drive, rq, blk_rq_pos(rq)); in start_request()
Dide-cd.c727 (blk_rq_pos(rq) & (sectors_per_frame - 1))) in cdrom_start_rw()
1313 long block = (long)blk_rq_pos(rq) / (hard_sect >> 9); in ide_cdrom_prep_fs()
Dide-tape.c576 rq->cmd[0], (unsigned long long)blk_rq_pos(rq), in idetape_do_request()
/linux-4.1.27/drivers/staging/i2o/
Di2o_block.c841 *((u32 *) & cmd[2]) = cpu_to_be32(blk_rq_pos(req) * hwsec); in i2o_block_transfer()
854 cpu_to_le32((u32) (blk_rq_pos(req) << KERNEL_SECTOR_SHIFT)); in i2o_block_transfer()
856 cpu_to_le32(blk_rq_pos(req) >> (32 - KERNEL_SECTOR_SHIFT)); in i2o_block_transfer()
/linux-4.1.27/drivers/mtd/
Dmtd_blkdevs.c83 block = blk_rq_pos(req) << 9 >> tr->blkshift; in do_blktrans_request()
93 if (blk_rq_pos(req) + blk_rq_cur_sectors(req) > in do_blktrans_request()
/linux-4.1.27/drivers/mmc/card/
Dblock.c1058 from = blk_rq_pos(req); in mmc_blk_issue_discard_rq()
1102 from = blk_rq_pos(req); in mmc_blk_issue_secdiscard_rq()
1278 (unsigned)blk_rq_pos(req), in mmc_blk_err_check()
1384 brq->cmd.arg = blk_rq_pos(req); in mmc_blk_rw_rq_prep()
1689 blk_rq_pos(prq) : blk_rq_pos(prq) << 9; in mmc_blk_packed_hdr_wrq_prep()
1705 brq->cmd.arg = blk_rq_pos(req); in mmc_blk_packed_hdr_wrq_prep()
/linux-4.1.27/drivers/sbus/char/
Djsflash.c193 unsigned long offset = blk_rq_pos(req) << 9; in jsfd_do_request()
/linux-4.1.27/kernel/trace/
Dblktrace.c726 __blk_add_trace(bt, blk_rq_pos(rq), nr_bytes, in blk_add_trace_rq()
957 __blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), in blk_add_trace_rq_remap()
986 __blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), 0, in blk_add_driver_data()
/linux-4.1.27/drivers/mtd/ubi/
Dblock.c197 pos = blk_rq_pos(req) << 9; in ubiblock_read()
/linux-4.1.27/drivers/cdrom/
Dgdrom.c603 block = blk_rq_pos(req)/GD_TO_BLK + GD_SESSION_OFFSET; in gdrom_readdisk_dma()
/linux-4.1.27/drivers/block/paride/
Dpd.c452 pd_block = blk_rq_pos(pd_req); in do_pd_io_start()
Dpcd.c748 pcd_sector = blk_rq_pos(pcd_req); in do_pcd_request()
Dpf.c788 pf_block = blk_rq_pos(pf_req); in do_pf_request()
/linux-4.1.27/drivers/md/
Ddm.c1986 blk_rq_pos(rq)); in map_request()
2108 pos = blk_rq_pos(rq); in dm_request_fn()
2719 pos = blk_rq_pos(rq); in dm_mq_queue_rq()
/linux-4.1.27/drivers/block/mtip32xx/
Dmtip32xx.c2385 u64 start = blk_rq_pos(rq); in mtip_hw_submit_io()
3721 err = mtip_send_trim(dd, blk_rq_pos(rq), blk_rq_sectors(rq)); in mtip_submit_request()
3748 if ((blk_rq_pos(rq) & 7) || (blk_rq_sectors(rq) & 7)) in mtip_check_unal_depth()
/linux-4.1.27/arch/um/drivers/
Dubd_kern.c1286 dev->rq_pos = blk_rq_pos(req); in do_ubd_request()
/linux-4.1.27/drivers/memstick/core/
Dmspro_block.c716 t_off = blk_rq_pos(msb->block_req); in mspro_block_issue_req()
Dms_block.c1909 lba = blk_rq_pos(msb->req); in msb_io_work()