Lines Matching refs:pending_req

261 					struct vscsibk_pend *pending_req)  in scsiback_print_status()  argument
263 struct scsiback_tpg *tpg = pending_req->v2p->tpg; in scsiback_print_status()
266 tpg->tport->tport_name, pending_req->v2p->lun, in scsiback_print_status()
267 pending_req->cmnd[0], status_byte(errors), msg_byte(errors), in scsiback_print_status()
325 uint32_t resid, struct vscsibk_pend *pending_req) in scsiback_do_resp_with_sense() argument
328 struct vscsibk_info *info = pending_req->info; in scsiback_do_resp_with_sense()
340 ring_res->rqid = pending_req->rqid; in scsiback_do_resp_with_sense()
361 if (pending_req->v2p) in scsiback_do_resp_with_sense()
362 kref_put(&pending_req->v2p->kref, in scsiback_do_resp_with_sense()
366 static void scsiback_cmd_done(struct vscsibk_pend *pending_req) in scsiback_cmd_done() argument
368 struct vscsibk_info *info = pending_req->info; in scsiback_cmd_done()
373 sense_buffer = pending_req->sense_buffer; in scsiback_cmd_done()
374 resid = pending_req->se_cmd.residual_count; in scsiback_cmd_done()
375 errors = pending_req->result; in scsiback_cmd_done()
378 scsiback_print_status(sense_buffer, errors, pending_req); in scsiback_cmd_done()
380 scsiback_fast_flush_area(pending_req); in scsiback_cmd_done()
381 scsiback_do_resp_with_sense(sense_buffer, errors, resid, pending_req); in scsiback_cmd_done()
385 static void scsiback_cmd_exec(struct vscsibk_pend *pending_req) in scsiback_cmd_exec() argument
387 struct se_cmd *se_cmd = &pending_req->se_cmd; in scsiback_cmd_exec()
388 struct se_session *sess = pending_req->v2p->tpg->tpg_nexus->tvn_se_sess; in scsiback_cmd_exec()
391 memset(pending_req->sense_buffer, 0, VSCSIIF_SENSE_BUFFERSIZE); in scsiback_cmd_exec()
395 scsiback_get(pending_req->info); in scsiback_cmd_exec()
396 se_cmd->tag = pending_req->rqid; in scsiback_cmd_exec()
397 rc = target_submit_cmd_map_sgls(se_cmd, sess, pending_req->cmnd, in scsiback_cmd_exec()
398 pending_req->sense_buffer, pending_req->v2p->lun, in scsiback_cmd_exec()
399 pending_req->data_len, 0, in scsiback_cmd_exec()
400 pending_req->sc_data_direction, 0, in scsiback_cmd_exec()
401 pending_req->sgl, pending_req->n_sg, in scsiback_cmd_exec()
433 static int scsiback_gnttab_data_map_list(struct vscsibk_pend *pending_req, in scsiback_gnttab_data_map_list() argument
439 struct vscsibk_info *info = pending_req->info; in scsiback_gnttab_data_map_list()
455 pending_req->n_grants += mapcount; in scsiback_gnttab_data_map_list()
461 pending_req->n_grants += mapcount; in scsiback_gnttab_data_map_list()
466 struct vscsibk_pend *pending_req) in scsiback_gnttab_data_map() argument
478 pending_req->n_sg = 0; in scsiback_gnttab_data_map()
479 pending_req->n_grants = 0; in scsiback_gnttab_data_map()
480 pending_req->data_len = 0; in scsiback_gnttab_data_map()
493 err = scsiback_gnttab_data_map_list(pending_req, ring_req->seg, in scsiback_gnttab_data_map()
494 pending_req->pages, pending_req->grant_handles, in scsiback_gnttab_data_map()
517 pending_req->sgl = kmalloc_array(nr_segments, in scsiback_gnttab_data_map()
519 if (!pending_req->sgl) in scsiback_gnttab_data_map()
522 sg_init_table(pending_req->sgl, nr_segments); in scsiback_gnttab_data_map()
523 pending_req->n_sg = nr_segments; in scsiback_gnttab_data_map()
526 if (pending_req->sc_data_direction == DMA_TO_DEVICE) in scsiback_gnttab_data_map()
529 pg = pending_req->pages + nr_sgl; in scsiback_gnttab_data_map()
530 grant = pending_req->grant_handles + nr_sgl; in scsiback_gnttab_data_map()
533 err = scsiback_gnttab_data_map_list(pending_req, seg, in scsiback_gnttab_data_map()
540 vaddr(pending_req, i) + ring_req->seg[i].offset); in scsiback_gnttab_data_map()
543 err = scsiback_gnttab_data_map_list(pending_req, seg, in scsiback_gnttab_data_map()
550 end_seg = vaddr(pending_req, 0) + ring_req->seg[0].offset; in scsiback_gnttab_data_map()
553 pg = pending_req->pages + nr_sgl; in scsiback_gnttab_data_map()
556 for_each_sg(pending_req->sgl, sg, nr_segments, i) { in scsiback_gnttab_data_map()
558 pending_req->data_len += seg->length; in scsiback_gnttab_data_map()
562 end_seg = vaddr(pending_req, i_seg) + in scsiback_gnttab_data_map()
586 static void scsiback_device_action(struct vscsibk_pend *pending_req, in scsiback_device_action() argument
590 struct scsiback_tpg *tpg = pending_req->v2p->tpg; in scsiback_device_action()
591 struct se_cmd *se_cmd = &pending_req->se_cmd; in scsiback_device_action()
602 &pending_req->sense_buffer[0]); in scsiback_device_action()
610 if (transport_lookup_tmr_lun(se_cmd, pending_req->v2p->lun) < 0) in scsiback_device_action()
621 transport_generic_free_cmd(&pending_req->se_cmd, 1); in scsiback_device_action()
625 scsiback_do_resp_with_sense(NULL, err, 0, pending_req); in scsiback_device_action()
627 kmem_cache_free(scsiback_cachep, pending_req); in scsiback_device_action()
658 struct vscsibk_pend *pending_req) in prepare_pending_reqs() argument
663 pending_req->rqid = ring_req->rqid; in prepare_pending_reqs()
664 pending_req->info = info; in prepare_pending_reqs()
672 pending_req->v2p = NULL; in prepare_pending_reqs()
677 pending_req->v2p = v2p; in prepare_pending_reqs()
680 pending_req->sc_data_direction = ring_req->sc_data_direction; in prepare_pending_reqs()
681 if ((pending_req->sc_data_direction != DMA_BIDIRECTIONAL) && in prepare_pending_reqs()
682 (pending_req->sc_data_direction != DMA_TO_DEVICE) && in prepare_pending_reqs()
683 (pending_req->sc_data_direction != DMA_FROM_DEVICE) && in prepare_pending_reqs()
684 (pending_req->sc_data_direction != DMA_NONE)) { in prepare_pending_reqs()
686 pending_req->sc_data_direction); in prepare_pending_reqs()
690 pending_req->cmd_len = ring_req->cmd_len; in prepare_pending_reqs()
691 if (pending_req->cmd_len > VSCSIIF_MAX_COMMAND_SIZE) { in prepare_pending_reqs()
693 pending_req->cmd_len); in prepare_pending_reqs()
696 memcpy(pending_req->cmnd, ring_req->cmnd, pending_req->cmd_len); in prepare_pending_reqs()
705 struct vscsibk_pend *pending_req; in scsiback_do_cmd_fn() local
725 pending_req = kmem_cache_alloc(scsiback_cachep, GFP_KERNEL); in scsiback_do_cmd_fn()
726 if (!pending_req) in scsiback_do_cmd_fn()
732 err = prepare_pending_reqs(info, &ring_req, pending_req); in scsiback_do_cmd_fn()
743 pending_req); in scsiback_do_cmd_fn()
744 kmem_cache_free(scsiback_cachep, pending_req); in scsiback_do_cmd_fn()
750 if (scsiback_gnttab_data_map(&ring_req, pending_req)) { in scsiback_do_cmd_fn()
751 scsiback_fast_flush_area(pending_req); in scsiback_do_cmd_fn()
753 DRIVER_ERROR << 24, 0, pending_req); in scsiback_do_cmd_fn()
754 kmem_cache_free(scsiback_cachep, pending_req); in scsiback_do_cmd_fn()
756 scsiback_cmd_exec(pending_req); in scsiback_do_cmd_fn()
760 scsiback_device_action(pending_req, TMR_ABORT_TASK, in scsiback_do_cmd_fn()
764 scsiback_device_action(pending_req, TMR_LUN_RESET, 0); in scsiback_do_cmd_fn()
769 0, pending_req); in scsiback_do_cmd_fn()
770 kmem_cache_free(scsiback_cachep, pending_req); in scsiback_do_cmd_fn()
1358 struct vscsibk_pend *pending_req = container_of(se_cmd, in scsiback_release_cmd() local
1361 kmem_cache_free(scsiback_cachep, pending_req); in scsiback_release_cmd()
1402 struct vscsibk_pend *pending_req = container_of(se_cmd, in scsiback_queue_data_in() local
1405 pending_req->result = SAM_STAT_GOOD; in scsiback_queue_data_in()
1406 scsiback_cmd_done(pending_req); in scsiback_queue_data_in()
1412 struct vscsibk_pend *pending_req = container_of(se_cmd, in scsiback_queue_status() local
1418 pending_req->result = (DRIVER_SENSE << 24) | in scsiback_queue_status()
1421 pending_req->result = se_cmd->scsi_status; in scsiback_queue_status()
1423 scsiback_cmd_done(pending_req); in scsiback_queue_status()