tgtport 36 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport; tgtport 79 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport; tgtport 105 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport; tgtport 142 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport; tgtport 153 drivers/nvme/target/fc.c return (iodptr - iodptr->tgtport->iod); tgtport 225 drivers/nvme/target/fc.c static void nvmet_fc_tgtport_put(struct nvmet_fc_tgtport *tgtport); tgtport 226 drivers/nvme/target/fc.c static int nvmet_fc_tgtport_get(struct nvmet_fc_tgtport *tgtport); tgtport 227 drivers/nvme/target/fc.c static void nvmet_fc_handle_fcp_rqst(struct nvmet_fc_tgtport *tgtport, tgtport 325 drivers/nvme/target/fc.c nvmet_fc_alloc_ls_iodlist(struct nvmet_fc_tgtport *tgtport) tgtport 335 drivers/nvme/target/fc.c tgtport->iod = iod; tgtport 339 drivers/nvme/target/fc.c iod->tgtport = tgtport; tgtport 340 drivers/nvme/target/fc.c list_add_tail(&iod->ls_list, &tgtport->ls_list); tgtport 349 drivers/nvme/target/fc.c iod->rspdma = fc_dma_map_single(tgtport->dev, iod->rspbuf, tgtport 352 drivers/nvme/target/fc.c if (fc_dma_mapping_error(tgtport->dev, iod->rspdma)) tgtport 362 drivers/nvme/target/fc.c fc_dma_unmap_single(tgtport->dev, iod->rspdma, tgtport 374 drivers/nvme/target/fc.c nvmet_fc_free_ls_iodlist(struct nvmet_fc_tgtport *tgtport) tgtport 376 drivers/nvme/target/fc.c struct nvmet_fc_ls_iod *iod = tgtport->iod; tgtport 380 drivers/nvme/target/fc.c fc_dma_unmap_single(tgtport->dev, tgtport 386 drivers/nvme/target/fc.c kfree(tgtport->iod); tgtport 390 drivers/nvme/target/fc.c nvmet_fc_alloc_ls_iod(struct nvmet_fc_tgtport *tgtport) tgtport 395 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 396 drivers/nvme/target/fc.c iod = list_first_entry_or_null(&tgtport->ls_list, tgtport 399 drivers/nvme/target/fc.c list_move_tail(&iod->ls_list, &tgtport->ls_busylist); tgtport 400 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 406 drivers/nvme/target/fc.c nvmet_fc_free_ls_iod(struct nvmet_fc_tgtport *tgtport, tgtport 411 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 412 drivers/nvme/target/fc.c list_move(&iod->ls_list, &tgtport->ls_list); tgtport 413 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 417 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_iodlist(struct nvmet_fc_tgtport *tgtport, tgtport 425 drivers/nvme/target/fc.c fod->tgtport = tgtport; tgtport 434 drivers/nvme/target/fc.c fod->rspdma = fc_dma_map_single(tgtport->dev, &fod->rspiubuf, tgtport 436 drivers/nvme/target/fc.c if (fc_dma_mapping_error(tgtport->dev, fod->rspdma)) { tgtport 439 drivers/nvme/target/fc.c fc_dma_unmap_single(tgtport->dev, fod->rspdma, tgtport 452 drivers/nvme/target/fc.c nvmet_fc_destroy_fcp_iodlist(struct nvmet_fc_tgtport *tgtport, tgtport 460 drivers/nvme/target/fc.c fc_dma_unmap_single(tgtport->dev, fod->rspdma, tgtport 488 drivers/nvme/target/fc.c nvmet_fc_queue_fcp_req(struct nvmet_fc_tgtport *tgtport, tgtport 499 drivers/nvme/target/fc.c ((queue->qid - 1) % tgtport->ops->max_hw_queues) : 0; tgtport 501 drivers/nvme/target/fc.c nvmet_fc_handle_fcp_rqst(tgtport, fod); tgtport 511 drivers/nvme/target/fc.c nvmet_fc_queue_fcp_req(fod->tgtport, fod->queue, fod->fcpreq); tgtport 520 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = fod->tgtport; tgtport 524 drivers/nvme/target/fc.c fc_dma_sync_single_for_cpu(tgtport->dev, fod->rspdma, tgtport 535 drivers/nvme/target/fc.c tgtport->ops->fcp_req_release(&tgtport->fc_target_port, fcpreq); tgtport 570 drivers/nvme/target/fc.c tgtport->ops->defer_rcv(&tgtport->fc_target_port, fcpreq); tgtport 599 drivers/nvme/target/fc.c assoc->tgtport->fc_target_port.port_num, tgtport 617 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_iodlist(assoc->tgtport, queue); tgtport 624 drivers/nvme/target/fc.c spin_lock_irqsave(&assoc->tgtport->lock, flags); tgtport 626 drivers/nvme/target/fc.c spin_unlock_irqrestore(&assoc->tgtport->lock, flags); tgtport 631 drivers/nvme/target/fc.c nvmet_fc_destroy_fcp_iodlist(assoc->tgtport, queue); tgtport 648 drivers/nvme/target/fc.c spin_lock_irqsave(&queue->assoc->tgtport->lock, flags); tgtport 650 drivers/nvme/target/fc.c spin_unlock_irqrestore(&queue->assoc->tgtport->lock, flags); tgtport 652 drivers/nvme/target/fc.c nvmet_fc_destroy_fcp_iodlist(queue->assoc->tgtport, queue); tgtport 677 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = queue->assoc->tgtport; tgtport 703 drivers/nvme/target/fc.c tgtport->ops->fcp_abort( tgtport 704 drivers/nvme/target/fc.c &tgtport->fc_target_port, fod->fcpreq); tgtport 725 drivers/nvme/target/fc.c tgtport->ops->defer_rcv(&tgtport->fc_target_port, tgtport 728 drivers/nvme/target/fc.c tgtport->ops->fcp_abort(&tgtport->fc_target_port, tgtport 731 drivers/nvme/target/fc.c tgtport->ops->fcp_req_release(&tgtport->fc_target_port, tgtport 752 drivers/nvme/target/fc.c nvmet_fc_find_target_queue(struct nvmet_fc_tgtport *tgtport, tgtport 764 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 765 drivers/nvme/target/fc.c list_for_each_entry(assoc, &tgtport->assoc_list, a_list) { tgtport 772 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 776 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 791 drivers/nvme/target/fc.c nvmet_fc_alloc_target_assoc(struct nvmet_fc_tgtport *tgtport) tgtport 803 drivers/nvme/target/fc.c idx = ida_simple_get(&tgtport->assoc_cnt, 0, 0, GFP_KERNEL); tgtport 807 drivers/nvme/target/fc.c if (!nvmet_fc_tgtport_get(tgtport)) tgtport 810 drivers/nvme/target/fc.c assoc->tgtport = tgtport; tgtport 820 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 822 drivers/nvme/target/fc.c list_for_each_entry(tmpassoc, &tgtport->assoc_list, a_list) tgtport 829 drivers/nvme/target/fc.c list_add_tail(&assoc->a_list, &tgtport->assoc_list); tgtport 831 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 837 drivers/nvme/target/fc.c ida_simple_remove(&tgtport->assoc_cnt, idx); tgtport 848 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = assoc->tgtport; tgtport 851 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 853 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 854 drivers/nvme/target/fc.c ida_simple_remove(&tgtport->assoc_cnt, assoc->a_id); tgtport 856 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(tgtport); tgtport 874 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = assoc->tgtport; tgtport 879 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 885 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 888 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 891 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 897 drivers/nvme/target/fc.c nvmet_fc_find_target_assoc(struct nvmet_fc_tgtport *tgtport, tgtport 904 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 905 drivers/nvme/target/fc.c list_for_each_entry(assoc, &tgtport->assoc_list, a_list) { tgtport 912 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 918 drivers/nvme/target/fc.c nvmet_fc_portentry_bind(struct nvmet_fc_tgtport *tgtport, tgtport 924 drivers/nvme/target/fc.c pe->tgtport = tgtport; tgtport 925 drivers/nvme/target/fc.c tgtport->pe = pe; tgtport 930 drivers/nvme/target/fc.c pe->node_name = tgtport->fc_target_port.node_name; tgtport 931 drivers/nvme/target/fc.c pe->port_name = tgtport->fc_target_port.port_name; tgtport 943 drivers/nvme/target/fc.c if (pe->tgtport) tgtport 944 drivers/nvme/target/fc.c pe->tgtport->pe = NULL; tgtport 955 drivers/nvme/target/fc.c nvmet_fc_portentry_unbind_tgt(struct nvmet_fc_tgtport *tgtport) tgtport 961 drivers/nvme/target/fc.c pe = tgtport->pe; tgtport 963 drivers/nvme/target/fc.c pe->tgtport = NULL; tgtport 964 drivers/nvme/target/fc.c tgtport->pe = NULL; tgtport 977 drivers/nvme/target/fc.c nvmet_fc_portentry_rebind_tgt(struct nvmet_fc_tgtport *tgtport) tgtport 984 drivers/nvme/target/fc.c if (tgtport->fc_target_port.node_name == pe->node_name && tgtport 985 drivers/nvme/target/fc.c tgtport->fc_target_port.port_name == pe->port_name) { tgtport 986 drivers/nvme/target/fc.c WARN_ON(pe->tgtport); tgtport 987 drivers/nvme/target/fc.c tgtport->pe = pe; tgtport 988 drivers/nvme/target/fc.c pe->tgtport = tgtport; tgtport 1096 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = tgtport 1098 drivers/nvme/target/fc.c struct device *dev = tgtport->dev; tgtport 1102 drivers/nvme/target/fc.c list_del(&tgtport->tgt_list); tgtport 1105 drivers/nvme/target/fc.c nvmet_fc_free_ls_iodlist(tgtport); tgtport 1108 drivers/nvme/target/fc.c tgtport->ops->targetport_delete(&tgtport->fc_target_port); tgtport 1111 drivers/nvme/target/fc.c tgtport->fc_target_port.port_num); tgtport 1113 drivers/nvme/target/fc.c ida_destroy(&tgtport->assoc_cnt); tgtport 1115 drivers/nvme/target/fc.c kfree(tgtport); tgtport 1121 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(struct nvmet_fc_tgtport *tgtport) tgtport 1123 drivers/nvme/target/fc.c kref_put(&tgtport->ref, nvmet_fc_free_tgtport); tgtport 1127 drivers/nvme/target/fc.c nvmet_fc_tgtport_get(struct nvmet_fc_tgtport *tgtport) tgtport 1129 drivers/nvme/target/fc.c return kref_get_unless_zero(&tgtport->ref); tgtport 1133 drivers/nvme/target/fc.c __nvmet_fc_free_assocs(struct nvmet_fc_tgtport *tgtport) tgtport 1138 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 1140 drivers/nvme/target/fc.c &tgtport->assoc_list, a_list) { tgtport 1146 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 1155 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport, *next; tgtport 1163 drivers/nvme/target/fc.c list_for_each_entry_safe(tgtport, next, &nvmet_fc_target_list, tgtport 1165 drivers/nvme/target/fc.c if (!nvmet_fc_tgtport_get(tgtport)) tgtport 1169 drivers/nvme/target/fc.c spin_lock_irqsave(&tgtport->lock, flags); tgtport 1170 drivers/nvme/target/fc.c list_for_each_entry(assoc, &tgtport->assoc_list, a_list) { tgtport 1178 drivers/nvme/target/fc.c spin_unlock_irqrestore(&tgtport->lock, flags); tgtport 1180 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(tgtport); tgtport 1207 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = targetport_to_tgtport(target_port); tgtport 1209 drivers/nvme/target/fc.c nvmet_fc_portentry_unbind_tgt(tgtport); tgtport 1212 drivers/nvme/target/fc.c __nvmet_fc_free_assocs(tgtport); tgtport 1214 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(tgtport); tgtport 1321 drivers/nvme/target/fc.c nvmet_fc_ls_create_association(struct nvmet_fc_tgtport *tgtport, tgtport 1359 drivers/nvme/target/fc.c iod->assoc = nvmet_fc_alloc_target_assoc(tgtport); tgtport 1371 drivers/nvme/target/fc.c dev_err(tgtport->dev, tgtport 1407 drivers/nvme/target/fc.c nvmet_fc_ls_create_connection(struct nvmet_fc_tgtport *tgtport, tgtport 1445 drivers/nvme/target/fc.c iod->assoc = nvmet_fc_find_target_assoc(tgtport, tgtport 1462 drivers/nvme/target/fc.c dev_err(tgtport->dev, tgtport 1495 drivers/nvme/target/fc.c nvmet_fc_ls_disconnect(struct nvmet_fc_tgtport *tgtport, tgtport 1531 drivers/nvme/target/fc.c assoc = nvmet_fc_find_target_assoc(tgtport, tgtport 1539 drivers/nvme/target/fc.c dev_err(tgtport->dev, tgtport 1580 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = iod->tgtport; tgtport 1582 drivers/nvme/target/fc.c fc_dma_sync_single_for_cpu(tgtport->dev, iod->rspdma, tgtport 1584 drivers/nvme/target/fc.c nvmet_fc_free_ls_iod(tgtport, iod); tgtport 1585 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(tgtport); tgtport 1589 drivers/nvme/target/fc.c nvmet_fc_xmt_ls_rsp(struct nvmet_fc_tgtport *tgtport, tgtport 1594 drivers/nvme/target/fc.c fc_dma_sync_single_for_device(tgtport->dev, iod->rspdma, tgtport 1597 drivers/nvme/target/fc.c ret = tgtport->ops->xmt_ls_rsp(&tgtport->fc_target_port, iod->lsreq); tgtport 1606 drivers/nvme/target/fc.c nvmet_fc_handle_ls_rqst(struct nvmet_fc_tgtport *tgtport, tgtport 1629 drivers/nvme/target/fc.c nvmet_fc_ls_create_association(tgtport, iod); tgtport 1633 drivers/nvme/target/fc.c nvmet_fc_ls_create_connection(tgtport, iod); tgtport 1637 drivers/nvme/target/fc.c nvmet_fc_ls_disconnect(tgtport, iod); tgtport 1645 drivers/nvme/target/fc.c nvmet_fc_xmt_ls_rsp(tgtport, iod); tgtport 1656 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = iod->tgtport; tgtport 1658 drivers/nvme/target/fc.c nvmet_fc_handle_ls_rqst(tgtport, iod); tgtport 1684 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = targetport_to_tgtport(target_port); tgtport 1690 drivers/nvme/target/fc.c if (!nvmet_fc_tgtport_get(tgtport)) tgtport 1693 drivers/nvme/target/fc.c iod = nvmet_fc_alloc_ls_iod(tgtport); tgtport 1695 drivers/nvme/target/fc.c nvmet_fc_tgtport_put(tgtport); tgtport 1729 drivers/nvme/target/fc.c fod->data_sg_cnt = fc_dma_map_sg(fod->tgtport->dev, sg, nent, tgtport 1747 drivers/nvme/target/fc.c fc_dma_unmap_sg(fod->tgtport->dev, fod->data_sg, fod->data_sg_cnt, tgtport 1773 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_rsp(struct nvmet_fc_tgtport *tgtport, tgtport 1831 drivers/nvme/target/fc.c fc_dma_sync_single_for_device(tgtport->dev, fod->rspdma, tgtport 1838 drivers/nvme/target/fc.c nvmet_fc_abort_op(struct nvmet_fc_tgtport *tgtport, tgtport 1852 drivers/nvme/target/fc.c tgtport->ops->fcp_abort(&tgtport->fc_target_port, fcpreq); tgtport 1858 drivers/nvme/target/fc.c nvmet_fc_xmt_fcp_rsp(struct nvmet_fc_tgtport *tgtport, tgtport 1866 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_rsp(tgtport, fod); tgtport 1868 drivers/nvme/target/fc.c ret = tgtport->ops->fcp_op(&tgtport->fc_target_port, fod->fcpreq); tgtport 1870 drivers/nvme/target/fc.c nvmet_fc_abort_op(tgtport, fod); tgtport 1874 drivers/nvme/target/fc.c nvmet_fc_transfer_fcp_data(struct nvmet_fc_tgtport *tgtport, tgtport 1900 drivers/nvme/target/fc.c fcpreq->sg_cnt < tgtport->max_sg_cnt && tgtport 1927 drivers/nvme/target/fc.c (tgtport->ops->target_features & NVMET_FCTGTFEAT_READDATA_RSP)) { tgtport 1929 drivers/nvme/target/fc.c nvmet_fc_prep_fcp_rsp(tgtport, fod); tgtport 1932 drivers/nvme/target/fc.c ret = tgtport->ops->fcp_op(&tgtport->fc_target_port, fod->fcpreq); tgtport 1958 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = fod->tgtport; tgtport 1967 drivers/nvme/target/fc.c nvmet_fc_abort_op(tgtport, fod); tgtport 1981 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = fod->tgtport; tgtport 2012 drivers/nvme/target/fc.c nvmet_fc_transfer_fcp_data(tgtport, fod, tgtport 2027 drivers/nvme/target/fc.c nvmet_fc_abort_op(tgtport, fod); tgtport 2043 drivers/nvme/target/fc.c nvmet_fc_transfer_fcp_data(tgtport, fod, tgtport 2053 drivers/nvme/target/fc.c nvmet_fc_xmt_fcp_rsp(tgtport, fod); tgtport 2080 drivers/nvme/target/fc.c __nvmet_fc_fcp_nvme_cmd_done(struct nvmet_fc_tgtport *tgtport, tgtport 2097 drivers/nvme/target/fc.c nvmet_fc_abort_op(tgtport, fod); tgtport 2118 drivers/nvme/target/fc.c nvmet_fc_transfer_fcp_data(tgtport, fod, tgtport 2129 drivers/nvme/target/fc.c nvmet_fc_xmt_fcp_rsp(tgtport, fod); tgtport 2137 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = fod->tgtport; tgtport 2139 drivers/nvme/target/fc.c __nvmet_fc_fcp_nvme_cmd_done(tgtport, fod, 0); tgtport 2147 drivers/nvme/target/fc.c nvmet_fc_handle_fcp_rqst(struct nvmet_fc_tgtport *tgtport, tgtport 2158 drivers/nvme/target/fc.c if (!tgtport->pe) tgtport 2188 drivers/nvme/target/fc.c fod->req.port = tgtport->pe->port; tgtport 2224 drivers/nvme/target/fc.c nvmet_fc_transfer_fcp_data(tgtport, fod, NVMET_FCOP_WRITEDATA); tgtport 2238 drivers/nvme/target/fc.c nvmet_fc_abort_op(tgtport, fod); tgtport 2293 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = targetport_to_tgtport(target_port); tgtport 2307 drivers/nvme/target/fc.c queue = nvmet_fc_find_target_queue(tgtport, tgtport 2330 drivers/nvme/target/fc.c nvmet_fc_queue_fcp_req(tgtport, queue, fcpreq); tgtport 2335 drivers/nvme/target/fc.c if (!tgtport->ops->defer_rcv) { tgtport 2498 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport; tgtport 2522 drivers/nvme/target/fc.c list_for_each_entry(tgtport, &nvmet_fc_target_list, tgt_list) { tgtport 2523 drivers/nvme/target/fc.c if ((tgtport->fc_target_port.node_name == traddr.nn) && tgtport 2524 drivers/nvme/target/fc.c (tgtport->fc_target_port.port_name == traddr.pn)) { tgtport 2526 drivers/nvme/target/fc.c if (!tgtport->pe) { tgtport 2527 drivers/nvme/target/fc.c nvmet_fc_portentry_bind(tgtport, pe, port); tgtport 2556 drivers/nvme/target/fc.c struct nvmet_fc_tgtport *tgtport = pe->tgtport; tgtport 2558 drivers/nvme/target/fc.c if (tgtport && tgtport->ops->discovery_event) tgtport 2559 drivers/nvme/target/fc.c tgtport->ops->discovery_event(&tgtport->fc_target_port); tgtport 373 drivers/nvme/target/fcloop.c fcloop_tgt_discovery_evt(struct nvmet_fc_target_port *tgtport) tgtport 381 drivers/nvme/target/fcloop.c tgt_rscn->tport = tgtport->private; tgtport 615 drivers/nvme/target/fcloop.c fcloop_fcp_op(struct nvmet_fc_target_port *tgtport, tgtport 708 drivers/nvme/target/fcloop.c fcloop_tgt_fcp_abort(struct nvmet_fc_target_port *tgtport, tgtport 732 drivers/nvme/target/fcloop.c fcloop_fcp_req_release(struct nvmet_fc_target_port *tgtport, tgtport 840 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_ls_rsp(struct nvmet_fc_target_port *tgtport, tgtport 849 drivers/scsi/lpfc/lpfc_nvmet.c struct lpfc_nvmet_tgtport *nvmep = tgtport->private; tgtport 930 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_fcp_op(struct nvmet_fc_target_port *tgtport, tgtport 933 drivers/scsi/lpfc/lpfc_nvmet.c struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private; tgtport 1064 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_fcp_abort(struct nvmet_fc_target_port *tgtport, tgtport 1067 drivers/scsi/lpfc/lpfc_nvmet.c struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private; tgtport 1126 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_xmt_fcp_release(struct nvmet_fc_target_port *tgtport, tgtport 1129 drivers/scsi/lpfc/lpfc_nvmet.c struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private; tgtport 1169 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_defer_rcv(struct nvmet_fc_target_port *tgtport, tgtport 1203 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_discovery_event(struct nvmet_fc_target_port *tgtport) tgtport 1209 drivers/scsi/lpfc/lpfc_nvmet.c tgtp = tgtport->private; tgtport 831 include/linux/nvme-fc-driver.h void (*targetport_delete)(struct nvmet_fc_target_port *tgtport); tgtport 832 include/linux/nvme-fc-driver.h int (*xmt_ls_rsp)(struct nvmet_fc_target_port *tgtport, tgtport 834 include/linux/nvme-fc-driver.h int (*fcp_op)(struct nvmet_fc_target_port *tgtport, tgtport 836 include/linux/nvme-fc-driver.h void (*fcp_abort)(struct nvmet_fc_target_port *tgtport, tgtport 838 include/linux/nvme-fc-driver.h void (*fcp_req_release)(struct nvmet_fc_target_port *tgtport, tgtport 840 include/linux/nvme-fc-driver.h void (*defer_rcv)(struct nvmet_fc_target_port *tgtport, tgtport 842 include/linux/nvme-fc-driver.h void (*discovery_event)(struct nvmet_fc_target_port *tgtport); tgtport 860 include/linux/nvme-fc-driver.h int nvmet_fc_unregister_targetport(struct nvmet_fc_target_port *tgtport); tgtport 862 include/linux/nvme-fc-driver.h int nvmet_fc_rcv_ls_req(struct nvmet_fc_target_port *tgtport, tgtport 866 include/linux/nvme-fc-driver.h int nvmet_fc_rcv_fcp_req(struct nvmet_fc_target_port *tgtport, tgtport 870 include/linux/nvme-fc-driver.h void nvmet_fc_rcv_fcp_abort(struct nvmet_fc_target_port *tgtport,